WO2011016967A2 - Spatial interfaces for realtime networked communications - Google Patents

Spatial interfaces for realtime networked communications Download PDF

Info

Publication number
WO2011016967A2
WO2011016967A2 PCT/US2010/042119 US2010042119W WO2011016967A2 WO 2011016967 A2 WO2011016967 A2 WO 2011016967A2 US 2010042119 W US2010042119 W US 2010042119W WO 2011016967 A2 WO2011016967 A2 WO 2011016967A2
Authority
WO
WIPO (PCT)
Prior art keywords
communicants
virtual area
communicant
communication session
graphical representation
Prior art date
Application number
PCT/US2010/042119
Other languages
French (fr)
Other versions
WO2011016967A3 (en
Inventor
Matthew Leacock
David Van Wie
Original Assignee
Social Communications Company
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Social Communications Company filed Critical Social Communications Company
Priority to CN2010800346974A priority Critical patent/CN102483819A/en
Priority to EP10806818A priority patent/EP2460138A2/en
Publication of WO2011016967A2 publication Critical patent/WO2011016967A2/en
Publication of WO2011016967A3 publication Critical patent/WO2011016967A3/en
Priority to IL217290A priority patent/IL217290A0/en

Links

Classifications

    • G06Q50/40
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • H04L12/1813Arrangements for providing special services to substations for broadcast or conference, e.g. multicast for computer conferences, e.g. chat rooms
    • H04L12/1822Conducting the conference, e.g. admission, detection, selection or grouping of participants, correlating users to one or more conference sessions, prioritising transmission
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/131Protocols for games, networked simulations or virtual reality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/54Presence management, e.g. monitoring or registration for receipt of user log-on information, or the connection status of the users
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • H04L12/1813Arrangements for providing special services to substations for broadcast or conference, e.g. multicast for computer conferences, e.g. chat rooms
    • H04L12/1827Network arrangements for conference optimisation or adaptation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • H04L12/1813Arrangements for providing special services to substations for broadcast or conference, e.g. multicast for computer conferences, e.g. chat rooms
    • H04L12/1831Tracking arrangements for later retrieval, e.g. recording contents, participants activities or behavior, network status

Definitions

  • the invention features a method in accordance with which a current realtime communication session is established between communicants operating on respective network nodes A spatial visualization of the current realtime communication session is displayed.
  • the spatial visualization includes a graphical representation of each of the communicants in spatial relation to a graphical
  • each of the communication states corresponds to a state of a respective communication channel over which a respective one of the communicants is configured to communicate.
  • the invention features a method in accordance with which a current realtime communication session is established between communicants operating on respective network nodes.
  • a spatial visualization of the current realtime communication session is displayed.
  • the spatial visualization includes a graphical representation of each of the communicants in spatiai relation to a graphical representation of a virtual area.
  • a log of event descriptions is presented.
  • the event descriptions describe respective events involving interactions of the communicants in the virtual area.
  • the event descriptions are presented in contextua! association with elements of the spatial visualization of the current realtime communication session.
  • the invention features a method in accordance with which receipt of a command from a first communicant operating on a first network node to initiate a private communication with a second communicant operating on a second network node prompts a response that includes the following, A current realtime communication session is established between the first and second network nodes, A private virtual area associated with the first and second communicants is identified. Context configuration data associated with the private virtual area and generated in response to interactions of the first and second communicants in the private virtual area is retrieved, A spatial visualization of the current realtime communication session is displayed The spatial visualization includes graphical representations of the first and second communicants in spatial relation to a graphical representation of the virtual area configured in accordance with the context configuration data.
  • the invention also features apparatus operabie to implement the method described above and computer-readable media storing computer-readable instructions causing a computer to implement the method described above,
  • FiG 1 is a diagrammatic view of an embodiment of a network communication environment that includes a first client network node, a second client network node, and a synchronous conferencing server node.
  • FiG. 2 is a flow diagram of an embodiment of a method of visualizing realtime networked communications on a client network node.
  • FiGS. 3A-3D. 4. and 5 are diagrammatic views of spatial interfaces for realtime networked communications
  • FiG. 6 is a diagrammatic view of an embodiment of a spatial interface for realtime networked communications.
  • FiG. 7 is a flow diagram of an embodiment of a method of managing realtime networked communications.
  • FiG. 8 is a diagrammatic view of an embodiment of a spatial interface integrated with a realtime communications interface.
  • FIG. 9 is a diagrammatic view of an embodiment of the spatial interface shown in FlG, 8 integrated with an additional spatial interface,
  • FiG. 10 is a diagrammatic view of an embodiment of a graphical user interface.
  • FIG. 11 is a flow diagram of an embodiment of a method of managing realtime networked communications between networked communicants in a private virtual area.
  • FiG. 12 is a diagrammatic view of an embodiment of a process of generating a spatial visualization of a current realtime communication session.
  • PiG. 13 is a diagrammatic view of an embodiment of a data model relating area identifiers to communicants, template specifications, and context data.
  • [00171 FiG. 14 is a diagrammatic view of an embodiment of a data model relating interaction record identifiers with area identifiers and interaction records.
  • FlG. 15 is a diagrammatic view of an embodiment of a spatial interface integrated with a realtime communications interface for realtime networked
  • FiG. 16 is a diagrammatic view of an embodiment of the spatial interface shown in FiG. 15.
  • FIG. 17 is a diagrammatic view of an embodiment of a spatial interface integrated with a realtime communications interface for realtime networked
  • FiG 18 is a diagrammatic view of an embodiment of a network communication environment that includes a first client network node, a second client network node, and a virtual environment creator.
  • FiG. 19 is a block diagram of the network communication environment of FIG, 1 that shows components of an embodiment of a client network node
  • a "communicant” is a person who communicates or otherwise interacts with other persons over one or more network connections, where the communication or interaction may or may not occur in the context of a virtual area.
  • a "user” is a person who communicates or otherwise interacts with other persons over one or more network connections, where the communication or interaction may or may not occur in the context of a virtual area.
  • a "user” is a person who communicates or otherwise interacts with other persons over one or more network connections, where the communication or interaction may or may not occur in the context of a virtual area.
  • a "user” is a person who communicates or otherwise interacts with other persons over one or more network connections, where the communication or interaction may or may not occur in the context of a virtual area.
  • a "user” is a person who communicates or otherwise interacts with other persons over one or more network connections, where the communication or interaction may or may not occur in the context of a virtual area.
  • a "user” is a person who communicates or otherwise interacts with other persons over one or more
  • a 'realtime contact" of a user is a communicant or other person who has communicated with the user via a realtime communications platform.
  • a ⁇ computer is any machine, device, or apparatus that processes data according to computer-readable instructions that are stored on a computer-readable medium either temporariiy or permanently
  • a "computer operating system” is a software component of a computer system that manages and coordinates the performance of tasks and the sharing of computing and hardware resources
  • a "software application” (also referred to as software, an application, computer software, a computer application, a program, and a computer program ⁇ is a set of instructions that a computer can interpret and execute to perform one or more specific tasks.
  • a "computer data fiie 11 is a block of information that durably stores data for use by a software application.
  • a “window” is a visual area of a display that typically includes a user interface.
  • a window typically displays the output of a software process and typically enables a user to input commands or data for the software process.
  • a window that has a parent is called a "child window.”
  • a window that has no parent, or whose parent is the desktop window, is calied a "top-level window.”
  • a “desktop” is a system-defined window that paints the background of a graphical user interface (GUS) and serves as the base for all windows displayed by all software processes.
  • GUS graphical user interface
  • a "database '5 is an organized collection of records that are presented in a standardized format that can be searched by computers.
  • a database may be stored on a single computer-readable data storage medium on a single computer or it may be distributed across multiple computer-readable data storage media on one or more computers.
  • a “data sink” (referred to herein simply as a “sink”) is any of a device (e.g., a computer), part of a device, or software that receives data.
  • a “data source 5' (referred to herein simply as a “source”) is any of a device (e.g., a computer), part of a device, or software that originates data.
  • a ⁇ network node (also referred to simpiy as a "node”) is a junction or connection potnt in a communications network.
  • exemplary network nodes include, but are not limited to, a terminal a computer, and a network switch
  • a "server” network node is a host computer on a network that responds to requests for information or service.
  • a “client” network node is a computer on a network that requests information or service from a server.
  • a “network connection” is a link between two communicating network nodes.
  • the term local network node refers to a network node that currently is the primary subject of discussion.
  • a "connection handle” is a pointer or identifier ⁇ e.g., a uniform resource identifier (URi)) that can be used to establish a network connection with a communicant, resource, or service on a network node.
  • a "network communication” can include any type of information (e g , text, voice : audio, video, electronic mail message, data fil ⁇ ; motion data stream, and data packet) that is transmitted or otherwise conveyed from one network node to another network node over a network connection.
  • Synchronous conferencing refers to communications in which communicants participate at the same time. Synchronous conferencing encompasses all types of networked collaboration technologies, including instant messaging (e.g., text chat), audio conferencing, video conferencing, application sharing, and file sharing technologies.
  • a "communicant interaction 8 is any type of direct or indirect action or influence between a communicant and another network entity, which may include for example another communicant, a virtual area, or a network service.
  • exemplary types of communicant communications include communicants communicating with each other in realtime, a communicant entering a virtual area, and a communicant requesting access to a resource from a network service.
  • Presence refers to the ability and willingness of a networked entity (e.g., a communicant, service, or device) to communicate, where such willingness affects the ability to detect and obtain information about the state of the entity on a network and the ability to connect to the entity.
  • a "realtime data stream” is data that is structured and processed in a continuous flow and is designed to be received with no delay or oniy imperceptible delay.
  • Reaitime data streams include digital representations of voice, video, user movements, facial expressions and other physicai phenomena, as well as data within the computing environment that may benefit from rapid transmission, rapid execution, or both rapid transmission and rapid execution, including for example, avatar movement instructions, text chat, realtime data feeds (e.g., sensor data, machine control instructions, transaction streams and stock quote information feeds), and file transfers,
  • a "link” is a connection between two network nodes and represents the f ⁇ li bandwidth aliocated by the two nodes for real-time communication. Each link is divided into channeis that carry respective real-time data streams. Channels are aliocated to particuiar streams within the overall bandwidth that has been allocated to the link.
  • a "virtuai area” (aiso referred to as an "area” or a ⁇ k place”) is a representation of a computer-managed space or scene. Virtuai areas typically are one- dimensional, two-dimensional, or three-dimensional representations , although in some embodiments a virtual area may correspond to a single point. Oftentimes, a virtua! area is designed to simuiate a physical, real-world space. For example, using a traditional computer monitor, a virtual area may be visualized as a two-dimensional graphic of a three-dimensional computer-generated space. However, virtuai areas do not require an associated visualization to impiement switching rules.
  • a virtual area typically refers to an instance of a virtual area schema, where the schema defines the structure and contents of a virtuai area in terms of variables and the instance defines the structure and contents of a virtual area in terms of values that have been resolved from a particular context,
  • a “virtual area application” also referred to as a "virtual area
  • the virtual area application typically includes definitions of geometry, physics, and realtime switching ruies that are associated with one or more zones of the virtual area.
  • a “virtuai environment” is a representation of a computer-managed space that includes at least one virtual area and supports realtime communications between communicants.
  • a “zone” is a region of a virtual area that is associated with at least one switching rule or governance rule.
  • a “switching rule” is an instruction that specifies a connection or disconnection of one or more realtime data sources and one or more realtime data sinks subject to one or more conditions precedent, A switching rule controls switching (e.g., routing, connecting, and disconnecting) of realtime data streams between network nodes communicating tn the context of a virtual area.
  • a governance rule controls a communicant's access to a resource (e.g., an area, a region of an area, or the contents of that area or region), the scope of that access, and follow- on consequences of that access (e.g., a requirement that audit records relating to that access must be recorded).
  • a "renderable zone” is a zone that is associated with a respective visualization
  • a "position" in a virtual area refers to a location of a point or an area or a volume in the virtual area.
  • a point typically is represented by a single set of one- dimensional, two-dimensional, or three-dimensional coordinates (e.g., x, y, z) that define a spot in the virtual area.
  • An area typically is represented by the three-dimensional coordinates of three or more coplanar vertices that define a boundary of a closed two- dimensional shape in the virtual area
  • a volume typically is represented by the three- dimensional coordinates of four or more non-copianar vertices that define a closed boundary of a three-dimensional shape in the virtual area,
  • a "spatial state” is an attribute that describes where a user has presence in a virtual area.
  • the spatial state attribute typica ⁇ y has a respective vaiue (e.g., a zoneJD vaiue) for each of the zones in which the user has presence.
  • a "communication state 3 is an attribute that describes a state of a respective communication channel over which a respective one of the communicants is configured to communicate.
  • an "object” (also sometimes referred to as a "prop") is any type of discrete element in a virtual area that may be usefully treated separately from the geometry of the virtual area.
  • exemplary objects include doors, portals, windows, view screens, and speakerphone.
  • An object typically has attributes or properties that are separate and distinct from the attributes and properties of the virtual area.
  • An "avatar” is an object that represents a communicant in a virtual area.
  • the embodiments that are described herein provide improved systems and methods for visualizing reaitime network communications, in particular, these embodiments apply a spatial metaphor on top of reaitime networked communications.
  • the spatial metaphor provides a context for depicting the current communication states of the communicants involved in reaitime networked communications.
  • the spatial metaphor aiso provides a context for organizing the presentation of various interface elements that are used by communicants to participate in realtime networked
  • FIG. 1 shows an embodiment of an exemplary network
  • the communications environment 10 that inciudes a first client network node 12 (Client Node A), a second ciient network node 14 (Client Network Node B), and a synchronous conferencing server 16 that are interconnected by a network 18.
  • the first client network node 12 includes a computer-readable memory 20 ; a processor 22, and input/output (I/O) hardware 24 (including a display).
  • the processor 22 executes at least one communications appiication 26 that is stored in the memory 20.
  • the second client network node 14 typically is configured in substantially the same way as the first client network node 12. in some embodiments, the synchronous conferencing server 18 manages realtime communication sessions between the first and second client nodes 12, 14.
  • the network infrastructure service environment 30 aiso maintains a relationship database 36 that contains records 38 of interactions between communicants. Each interaction record 38 describes the context of an interaction between a pair of communicants.
  • the communications application 26 and the synchronous conferencing server 16 together provide a platform (referred to herein as "the platform") for creating a spatial visualization context that enhances realtime communications between communicants operating on the network nodes 12, 14.
  • FIG. 2 shows an embodiment of a method that is implemented by the communications application 26 operating on one or both of the first and second network nodes 12, 14. This process typically is performed in response to a request from a communicant on one of the network nodes 12, 14 to initiate a realtime communication session with another communicant operating on the other network node.
  • the communications appiication 26 establishes a current realtime communication session between communicants operating on respective network nodes (FIG. 2, block 40)
  • the communications application 26 dispiays a spatial visualization of the current realtime communication session (FSG. 2, biock 40).
  • the spatia! visualization includes a graphical representation of each of the communicants in spatial relation to a graphical representation of a virtual area.
  • the virtuai area may be represented graphically by any type of one-dimensional, two-dimensional, or three-dimensional view that situates the graphicai representations of the communicants in respective positions in a visual space.
  • the communications application 26 depicts visual cues in the spatiai visualization that show current communication states of the communicants (FlG, 2 S block 44). Each of the communication states typically
  • a state of a respective communication channel e.g., text chat, audio, video, appiication share, and file share channel
  • a respective communication channel e.g., text chat, audio, video, appiication share, and file share channel
  • a log of event descriptions that describe respective events involving interactions of the communicants in the virtual area is presented on the display in contextual association with elements of the spatial visualization of the current realtime communication session.
  • the log of event descriptions and the graphicai representation of the virtual area typically are displayed in a single graphicai user interface window.
  • the iog of event descriptions may include, for example, at least one of: text of a chat conversation between the communicants in the virtual area; a description of a data file shared by a respective one of the
  • a respective label is
  • the log of event descriptions typically are stored in one or more database records that are indexed by an identifier of the virtual area
  • one or more props are displayed in the virtual area, where each prop represents a respective communication channel for realtime communications between the communicants during the communication session.
  • a communicant-seiectabie table prop may be displayed in the virtual area, and a fiie share session between the communicants may be initiated in response to selection of the table prop by one of the communicants; or a communicant-selectable viewscreen prop may be displayed in the virtual area, and an application sharing session may be initiated between the communicants in response to selection of the viewscreen prop by one of the communicants.
  • a spatial property of the graphicai representation of a respective one of the communicants in relation to a respective one of the props is changed in response to selection of the respective prop by the respective communicant.
  • the graphical representation of the respective communicant may be depicted adjacent the selected prop, it may be reoriented to face the selected prop, and/or the graphical representation of the communicant may be changed (e.g., a pair of eyes may be added to the body of a communicant ' s sprite when it is positioned adjacent to a viewscreen prop, as shown in FIGS. 15 and 16).
  • a realtime instant messaging communication channel is established between the communicants during the current communication session, in these embodiments, a current chat log of a current chat conversation between the communicants occurring during the current communication session typically is displayed in association with the graphical representation of the virtual area A respective prior chat log of a prior chat conversation that occurred during a prior communication session between the communicants in the virtual area typically is displayed in association with the current chat log.
  • the graphical representation of a given one of the communicants may be dynamically modulated in response to receipt of a respective realtime chat stream from the given communicant over the realtime instant messaging communication channel such that the current communication state of the given communicant is reflected in the dynamic modulation of the graphical
  • a graphical representation of a file sharing prop is displayed in the virtual area.
  • the graphical representation of the respective communtcant typically is depicted adjacent the file sharing prop and a realtime file sharing session typically is initiated in the virtual area,
  • a data file shared by the respective communicant during the realtime file sharing session typically is stored in a data storage device with an index that includes an identifier of the virtual area, and a communicant-selectable graphical representation of the data file typically is displayed on the file sharing prop.
  • a download of the data file to the network node from which a given one of the communicants is operating typically is initiated in response to selection of the graphical representation of the file by the given communicant,
  • a graphical representation of an application sharing prop is displayed in the virtual area.
  • the graphical representation of the respective communicant typically is depicted adjacent the application sharing prop and a realtime application sharing session typically is initiated in the virtual area.
  • a graphical indication that an application that is being shared typically is displayed in connection with the application sharing prop, in some embodiments, a first graphical representation of the application sharing prop is displayed during periods of application sharing between the communicants in the virtual area and a second graphical representation of the application sharing prop different from the first graphical representation is displayed during periods free of application sharing between the communicants.
  • a realtime audio communication channel is established between the given communicant and one or more of the other communicants configured as audio sources, and the depicting graphical representation of the given communicant is modified to show that the given communicant is configured as an audio sink.
  • a realtime audio communication channel is established between the given communicant and one or more of the other communicants configured as audio sinks, and the graphical representation of the given communicant is modified to show that the given communicant is configured as an audio source.
  • a static view of the graphica! representation of the virtual area is displayed throughout the current communication session, and the communicants are unable to navigate the graphical representations of the
  • the current realtime communication session between the first and second communicants is established, and the graphical representations of the first and second communicants are dispiayed in spatial relation to a graphica! representation of a virtual area that is indexed by identifiers of the first and second communicants.
  • an end state of a prior realtime communication session between the communicants is determined from data that is indexed by an identifier of the virtual area and that describes events that occurred during a prior communication session between the communicants, and the graphical representation of a virtual area is displayed in a state that corresponds to the determined end state of the prior communication session between the communicants.
  • FIGS. 3A-3D respectively show embodiments of spatial visualizations of a realtime communication session that include visual cues that reveal the current communication states of two networked communicants involved in the realtime communication session.
  • the spatial visualizations include a graphica! representation 46, 48 of each of the communicants in spatial relation to a graphical representation 50 of a virtual area.
  • the virtual area is
  • each communicant is represented by a respective circular sprite 48, 48.
  • the states of various communication channels over which the respective communicant is configured to communicate are revealed by visual cues that are shown in the spatial visualization. For example, the on or off state of a communicant's !oca! speaker channel is depicted by the presence or absence of a headphones graphic 52 on the communicant's sprite 46. Thus, when the speakers of the communicant who is represented by the sprite 46 are on, the headphones graphic 52 is present (as shown in FIG.
  • the on or off state of the communicant's microphone is depicted by the presence or absence of a microphone graphic 54 on the communicant's sprite 46 and a series of concentric circles 56 that radiate away from the communicant ' s sprite 46 in a series of expanding waves.
  • the microphone graphic 54 and the radiating concentric circles 58 are present (as shown in FIG. 3C) and, when the microphone is off, the microphone graphic 54 and the radiating concentric circles 56 are absent (as shown in FIGS. 3A, 3B, and 3D).
  • the headphones graphic 52, the microphone graphic 54, and the radiating concentric circles 56 serve as visuai cues of the states of the
  • communicant ' s sound playback and microphone devices The on or off state of a communicant's text chat channel is depicted by the presence or absence of a hand graphic 57 adjacent the communicant ' s sprite (as shown in FIG. 3D).
  • a communicant is transmitting text chat data to another network node the hand graphic 57 is present, and when a communicant is not transmitting text chat data the hand graphic 57 is not present.
  • text chat data is transmitted only when keyboard keys are depressed, in which case the visualization of the communicant s text channel appears as a flashing on and off of the hand graphic 57.
  • FIGS. 4 and 5 respectively show embodiments of spatial visualizations of a realtime communication session that include visua! cues that reveal the current communication states of two networked communicants involved in the realtime communication session in reiation to props (also referred to as objects) in a graphical representation of a virtual area.
  • the spatial visuaSization includes a graphical representation 46, 48 of each of the communicants in spatial relation to a graphical representation 58 of a virtual area.
  • the virtual area is represented by a perspective view of a three-dimensional visua! space in which the graphical representations 46, 48 of the communicants can have different respective positions
  • these visualizations include a viewscreen 60 thai shows the state of application sharing communication sessions, and a table 62 that shows the state of file sharing communication sessions.
  • the viewscreen 60 provides visual cues that indicate whether or not a communtcant is sharing an application over an application sharing channel. As shown in FIG. 4. in response to a communicant's selection of the viewscreen 60, the
  • communicant's sprite 48 automatically is moved to a position in the graphical
  • the position of the communicant's sprite 48 adjacent the viewscreen 60 indicates that the communicant currently is sharing or is about to share an application with the other communicants in the virtual area.
  • the graphical depiction of viewscreen 60 is changed depending on whether or not an active application sharing session is occurring, in the illustrated embodiment, the depicted color of the viewscreen 60 changes from light during an active application sharing session (as shown in FIG. 4) to dark when there is no application sharing taking place (as shown in FiG. 5). Additional details regarding the application sharing process are described in connection with FfGS. 26-28 of U. S Patent Application No. 12/354,709, filed January 15. 2009, and in U.S. Patent Application No. 12/418,270, filed April 3, 2009,
  • the table 62 provides visual cues that indicate whether or not a communicant is sharing or has shared a data file over a data file sharing channel. As shown in FiG. 5, in response to a communicant's selection of the table 62, the communicant's sprite 48 automatically is moved to a position in the graphical user interface.
  • the communicant uploads the data file from the client node 12 to a repository that is maintained by the synchronous conferencing server node 30.
  • the synchronous conferencing server node 30 stores the uploaded fiie in the repository and creates a database record that associates the data file with the table 62.
  • the state of the table 62 changes from having a clear table surface (as shown in FlG 4) to having a graphical representation 64 of a data file on the table surface (as shown in FiG. 5).
  • Other communicants in the virtual area 58 are able to view the contents of the uploaded data file by selecting the graphical representation 64 and, subject to governance rules associated with the virtual area 58, optionally may be able to modify or delete the data file. Additional details regarding the file sharing process are described in connection with FiGS, 22 and 23 of U.S. Patent Application No. 12/354,709, filed January 15, 2009.
  • FIG. 8 shows an embodiment of a spatia! visualization 70 of two realtime communication sessions in two different virtual areas (i.e., "Virtual Area I" and "Virtual Area Ii").
  • Each of the virtual areas is represented by a one-dimensional space that contains graphical representations of the communicants who currently have presence in the space.
  • the ordering of the spatial positions (e.g., from left to right) of the graphical representations of the communicants in each of the virtual areas corresponds to a spatial visualization of the temporal ordering of the communicants in terms of the times when they established respective presences in the virtual areas
  • each communicant is represented by a respective circular sprite 46, 48, 72, 74, 76, 78.
  • the communicant named "Dave" is represented by a respective sprite 48 : 78 in each of the virtual areas, reflecting the fact that he is present in both virtual areas.
  • the states of various communication channels over which the respective communicant is configured to communicate are revealed by visual cues that are shown in the spatial visualization 70.
  • the on or off state of a communicant's local speaker channel is depicted by the presence or absence of a headphones graphic 52 on the communicant ' s sprite.
  • the headphones graphic 52 is present (see sprites 46 ; 48 S 72,76, and 78 ⁇ and, when the communicant's speakers are off, the headphones graphic 52 is absent (see sprite 74).
  • the on or off state of the communicant's microphone is depicted by the presence or absence of a microphone graphic 54 on the communicant's sprite.
  • the headphones graphic 52 and the microphone graphic 54 provide visual cues of the states of the communicant's sound playback and microphone devices. iii. INTEGRATING SPATIAL VISUALIZATIONS WITH LOGS OF REALTIME NETWORKED INTERACTIONS IN A VIRTUAL AREA
  • Embodiments of the platform are capable of integrating a spatial visualization of realtime networked communications in a virtual area with logs of the interactions that are associated with the virtual area. In this way, current and prior Sogs of communicant interactions are enhanced with references to the spatiai visualization of those interactions, references which engage the communicants' spatial memories of the interactions to enable greater recall and understanding of the contexts of the
  • a current realtime communication session is established between communicants operating on respective network nodes.
  • a spatial visualization of the current realtime communication session is displayed on a display.
  • the spattal visualization includes a graphical representation of each of the
  • communicants in spatial relation to a graphical representation of a virtual area.
  • a log of event descriptions describing respective events involving interactions of the communicants in the virtual area is presented on the display in contextual association with elements of the spatial visualization of the current realtime communication session.
  • a visual association between respective ones of the event descriptions in the log and elements of the spatial visualization of the current realtime communication session is depicted on the display.
  • a visual association may be depicted between respective ones of the event descriptions in the log and respective ones of the graphical representations of the communicants involved in the events described by the respective event descriptions,
  • a respective label may be associated with each of one or more of the event
  • the label has a respective visual appearance that matches a visual element of the graphical representation of the communicant involved in the event described by the event description
  • the events in the log share a common visual vocabulary with the slate of the communicants in the spatiai visualization shown in the display.
  • the graphical representation of the respective communicant in response to an entry of a respective one of the communicants into the virtual area, is added to the spatial visualization, and a respective one of the event descriptions describing the entry of the respective communicant into the virtual area is presented on the display, in some embodiments, in response to a departure of a respective one of the communicants from the virt ⁇ ai area, the graphical representation of the respective communicant is removed from the spatsa! visualization, and a respective one of the event descriptions describing the departure of the respective communicant from the virtual area is presented on the display.
  • a communicant-seiectabie graphical representation of the data file is displayed in spatial relation to the graphical representation of the virtual area, and a respective one of the event descriptions describing the sharing of the data file by the respective communicant is presented on the display.
  • a graphical indication of the sharing of the application in spatial relation to the graphicai representation of the virtual area is displayed on the display, and a respective one of the event descriptions describing the sharing of the application by the respective communicant is displayed on the display.
  • FIG. 7 shows an embodiment of a method by which the platform integrates spatiai visualizations of realtime networked interactions in a virtual area with historical records of the interactions that are associated with the virtual area.
  • the platform retrieves context configuration data that includes a log of interactions that are associated with the virtual area (FlG, 7, block 82).
  • the tog typically includes data that is extracted from the interaction records 38, which describe the contexts of interactions between communicants in the virtual area.
  • the extracted data may include, for example, data stream data (e.g., text chat entries) and references (e.g., hyperlinks) to files an ⁇ data streams (e,g. : audio and video data streams) that are shared or recorded during one or more prior
  • the platform generates a visualization of the current realtime communication session in the virtual area in association with the historical log (FIG. 7, block 84).
  • the platform typically retrieves context data describing an end state of the preceding communications session in the virtual area : including the positions and states of the props in the virtual area.
  • the spatial visualization that is generated includes a graphical representation of each of the communicants in spatial relation to a graphical representation of the virtual area.
  • the virtual area may be represented graphically by any type of one-dimensional, two-dimensional, or three- dimensional view that situates the graphical representations of the communicants in respective positions in a visual space.
  • the platform depicts visual cues in the spatial visualization that shows current
  • Each of the communication states typically corresponds to a state of a respective communication channel (e.g., text chat, audio, video, application share, and file share channel) over which a respective one of the communicants is configured to communicate,
  • a respective communication channel e.g., text chat, audio, video, application share, and file share channel
  • the platform stores context configuration data that includes records of interactions between the
  • Each interaction record describes the context of an interaction between a pair of the communicants in the virtual area.
  • an interaction record contains an identifier for each of the communicants, an identifier for the place of interaction (e.g., a virtual area instance), a description of the hierarchy of the interaction place (e.g., a description of how the interaction area relates to a larger area), start and end times of the interaction, and a list of all files and other data streams that are shared or recorded during the interaction.
  • the interaction platform tracks when it occurred, where it occurred, and what happened during the interaction in terms of communicants involved (e g., entering and exiting), objects that are activated/deactivated, and the files that were shared.
  • the platform stores context configuration data that describes the end state of the current communication session (FIG. 7, block 90).
  • the en ⁇ state context configuration data typically includes a description of all props (e.g., viewscreen and table props) that are present in the virtual area at the time the current communication session was terminated, including a description of the positions of the props and their respective states (e.g., associations between a table prop and data files that were shared in the virtual area).
  • the en ⁇ state context configuration data typically is used by the platform to recreate the end slate of the virtual area for the next realtime communication session that takes place in the virtual area,
  • Some embodiments apply one or more of the spatial metaphor visualizations described above on top of realtime chat interactions. These visualizations provide a context for depicting the current communication states of the communicants involved in realtime chat interactions.
  • the spatial metaphor also provides a context for organizing the presentation of various interface elements that are used by
  • the spatial metaphor visualizations may be applied to any type of instant messaging platform that provides realtime text-based communication between two or more communicants over the internet or some form of interna! network/intranet, optionally with one or more other realtime communication channels, such as audio, video, file share, and application sharing channels.
  • embodiments may be integrated with any of the currently available instant messaging platforms including, for example, AOL Instant Messenger, MSN Messenger, Yahoo! Messenger, Google Talk, and Skype.
  • FIG. 8 shows an exemplary embodiment of a spatial interface 92 for a realtime chat interaction between a group of communicants in a virtual area.
  • Each of the communicants is represented graphically by a respective sprite 94, 98, 98, 100, 102 and the virtual area is represented graphically by a two-dimensional top view of a rectangular space 101 (i.e., the "West Conference" space).
  • thetr sprites automatically are positioned in predetermined locations (or "seats" in the virtual area.
  • the virtual area includes two viewscreen props 104, 106 and a table prop 108, Communicants interact with the props by selecting them with a input device (e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like).
  • a input device e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like.
  • the spatial interface 92 is integrated with a realtime communications interface window 110 that also includes a toolbar 112, a chat log area 114, a text box 1 16, and a Send button 118,
  • the user may enter text messages in the text box 116 and transmit the text messages to the other communicants in the currently West Conference space 101 by selecting the Send button 118.
  • the spatial interface 92 and the chat iog area 114 are separated by a splitter 117 that, in some embodiments, can be slid up and down by the user to hide or reveal the spatial interface 92.
  • the °haf log area 114 displays a log of current and optionally prior events that are associated with the West Conference space 101 .
  • An exemplary set of events that are dispiayed in the chat iog area 114 include: text messages that the user has exchanged with other communicants in the West Conference space 101 ; changes in the presence status of communicants in the West Conference space 101 ; changes in the speaker and microphone settings of the communicants in the West Conference space 101 ; and the status of the props 104-108, inciuding references to any
  • the events are labeled by the communicant's name followed by content associated with the event (e.g., a text message) or a description of the event.
  • content associated with the event e.g., a text message
  • status related events are labeled as follows;
  • each of the events is associated with a respective timestamp 119 that identifies the date and time when the associated event was initiated.
  • the chat log area 114 typically contains a standard "chat history" (also referred to as an "instant message history") that includes a list of entries typed remotely by two or more networked communicants, interleaved in the order the entries have been typed.
  • the chat history typically is displayed on each communicant's terminal display, along with an indication of which user made a particular entry and at what time relative to other communicant's entries. This provides a session history for the chat by enabling communicants to independently view the entries and the times at which each entry was made.
  • the spatial visualization 92 provides a context for organizing the presentation of the events that are displayed in the chat log area 114,
  • each of the displayed events is labeled with a respective tag that visually correlates with the appearance of the sprite of the communicant that sourced the displayed event.
  • each of the events that is sourced by a particular one of the communicants is labeled with a respective icon 130, 132, 134, 136 with a visual appearance ⁇ e.g., color-code) that matches the visual appearance of that communicant's sprite, in this example, the color of the icons 130, 134 matches the color of the body of Dave's sprite 100.
  • the color of the icon 132 matches the color of the body of Camilla's sprite 98
  • the color of the icon 136 matches the color of the body of Jack's sprite 96,
  • the toolbar 112 includes a set of navigation and interaction control buttons, including a headphones button 120 for toggling on and off the user's speakers, a microphone button 122 for toggling on and off the user's microphone, a get button 124 for getting people, a map button 126 for opening a map view of a larger virtual area the contains the space 101 , and a reconnect button 128 for reestablishing a connection to the virtual area.
  • a headphones button 120 for toggling on and off the user's speakers
  • a microphone button 122 for toggling on and off the user's microphone
  • a get button 124 for getting people
  • a map button 126 for opening a map view of a larger virtual area the contains the space 101
  • a reconnect button 128 for reestablishing a connection to the virtual area.
  • the user may toggle one or both of the headphones button 120 and the microphone button 122 in order to selectively turn-on and turn-off one or both of the user's speakers and microphone.
  • the headphones graphic, the radiating concentric circles around the user's sprite, and the microphone graphic on the user's sprite are omitted when the user's speakers and microphone both are turned-off.
  • a list of communicants is displayed in a separate frame 138.
  • the communicants are segmented into two groups: a first group labeled "People in West Conference" that identifies all the communicants who are in the current area (i.e., West Conference); and a second group labeled "Lansing Aviation” that identifies all the communicants who are present in a larger area (i.e., Lansing Aviation, which contains the current area) but are not present in the current area.
  • Each of the virtual areas is represented by a respective one-dimensional space 142, 144 that contains graphical representations of the communicants who currently have presence in the space.
  • each communicant is represented by a respective circular sprite that is labeled with a respective user name of the communicant (i.e., "Jack,” “Dave,” “Camilla:' “Karou/ ' “Arkadi/ 1 ⁇ uka,” “Teca/' ⁇ oshi/ ' and "Adam 1" ).
  • the headphones graphic 52 is present (see sprites Jack, Dave, Camilla, Karou, Arkadi, and Teca) and, when the communicant's speakers are off, the headphones graphic 52 is absent (see sprites Yuka : Yoshl and Adam).
  • the on or off state of the communicant's microphone is depicted by the presence or absence of a microphone graphic 54 on the communicant's sprite.
  • the microphone graphic 54 is present (see sprites Karou and Teca) and, when the microphone is off, the microphone graphic 54 is absent (see sprites Jack, Dave, CamiSla, Arkadi, Yuka, Yoshi, and Adam ⁇ .
  • the radiating circles that indicate the on state of a communicant's microphone graphic typically is omitted in this
  • the headphones graphic 52 and the microphone graphic 54 provide visual cues of the states of the communicant's sound playback and microphone devices.
  • the activity state of a communicant's text chat channel is depicted by the presence or absence of the hand graphic 57 adjacent the communicant's sprite (see sprite Adam).
  • text chat data is transmitted only when keyboard keys are depressed, in which case the visualization of the communicant's text channel appears as a flashing on and off of the hand graphic 57
  • the platform transmits an invitation to the selected communicant to join the user in the respective zone.
  • FIG. 10 shows a pop-up window 141 that is generated by the piatform in the situation in which the user has selected "Arkadi" in the iist of available communicants displayed in the frame 138.
  • the platform transmits an invitation to the communicant who is associated with the name Arkadi to join the user in the West Conference space 101 (e.g. , "Please join me in West Conference ⁇ Jack.” ⁇ .
  • Some embodiments apply one or more of the spatial metaphor visualizations described above on top of realtime private interactions between (typically only two) networked communicants.
  • These spatial visualizations enable the depiction of a current private realtime communications session between the communicants in the context of their prior private relationship history.
  • the semantics of the virtual area is the relationship history between the communicants.
  • the spatial visualizations also provide a framework for organizing the presentation of various interface elements that are used by communicants to participate in private realtime networked communications in the context of their prior relationship history.
  • communicants typically is visualized as a private virtual area that provides a reference for the records of the private interactions that occur in the private virtual area, records which are stored persistently in the relationship database 36 in association with the private virtual area.
  • the virtual area typically is created automatically during the first communication session and then persists until one or all of the communicants choose to delete it.
  • the private virtual area typically is owned jointly by all the
  • any of the communicants can freely access the private virtual area and the associated private interaction records, and can unilaterally add. copy, or delete the private virtual area and all the associated private interaction records.
  • Each communicant typically must explicitly navsgate to the private virtual area that he or she shares with another communicant, fn some embodiments, this is achieved by selecting an interface control that initiates a private communication with the other communicant.
  • a private instant messaging communication e.g., a text, audio, or video chat
  • the platform in response to the initiating of a private instant messaging communication (e.g., a text, audio, or video chat) with another communicant, the platform automatically situates the private communication in a private virtual area that typically is configured in accordance with configuration data that describes the prior state of the private virtual area when the communtcants last communicated in the private virtual area.
  • the platform responds to the receipt of a command from a first communicant operating on a first network node to initiate a private communication with a second communicant operating on a second network node as foliows.
  • the platform estabiishes a current realtime communication session between the first and second network nodes.
  • the piatform identifies a private virtual area that is associated with the first and second communicants,
  • the piatform retrieves context configuration data associated with the private virtual area and generated in response to interactions of the first and second communicants in the private virtual area.
  • the piatform displays a spatia ⁇ visualization of the current realtime
  • the spatial visualization includes graphical
  • the platform during the current realtime communication session, the platform generates a log of event descriptions describing respective events involving interactions of the first and second communicants in the virtual area.
  • piatform typically stores the event descriptions in a data storage device with an index comprising an identifier of the virtual area.
  • the log of event descriptions may include, for example, at least one of: text of a chat conversation between the first and second communicants in the virtual area; a description of a data file shared by a respective one of the first and second
  • communicants in the virtuai area typically presents the log of event descriptions on the display.
  • the log of event descriptions typically is presented in contextual association with elements of the spatial visualization of the current realtime
  • the platform retrieves context configuration data that includes a iog of event descriptions describing respective events involving interactions of the first and second communicants in the virtual area during one or more prior communication sessions before the current communication session.
  • the platform typicaily presents the tog of event descriptions generated during the current realtime communication session together with the retrieved context configuration data
  • the platform retrieves context configuration data that includes a description of an end state of a prior realtime communication session between the communicants and dispiays the graphical representation of a virtual area in a state that corresponds to the end state of the prior communication session between the communicants.
  • FiG. 11 shows an embodiment of a method of managing realtime networked communications between networked communicants in a private virtual area.
  • the platform determines whether or not a private virtual area that is indexed by the identifiers of all the communicants already has been created (FiG. 11 , block 152). If such a private virtual area already has been created, the platform retrieves a specification of the private virtual area (FlG, 11 ; block 154); the platform also, retrieves context configuration data that is associated with the private virtual area (FSG. 11 , block 156).
  • the platform creates a new private virtual area that is indexed by identifiers of ail the communicants (FSG. 11 , block 158).
  • the platform After the specification of the private virtual area has been either retrieved or newly created, the platform generates a visualization of the current realtime communication session in the private virtual area configured in its current context (i.e., either in its prior configuration or in its new default configuration) (FIG. 11 , block 160),
  • the platform stores context configuration data that describes the state of the private virtual area and includes records of interactions in the private virtual area, which records are indexed by the identifier of the private virtual area (FIG. 11 , block 162),
  • FIG. 12 shows an embodiment of a process 168 of generating a spatial visualization of a current realtime communication session.
  • each of the communicants (A and B) is represented by a respective node 170, 172 an ⁇ their private bilateral relationship ts represented by an edge 174 of a graph that interconnects the nodes 170, 172.
  • the bilateral relationship between the communicants is defined by their interaction history in the private virtual area.
  • the interaction history is stored in the interaction database 38 in the form of interaction records that describe the interactions of the communicants in the private virtual area. These interactions can include any of the interactions involving any of the communication channels over which the
  • communicants are configured to communicate, including, for example, chat, audio, video, realtime differential streams of tagged records containing configuration
  • the interaction history between the communicants is integrated with a template 178 that describes a graphical representation of the private virtual area to produce the spatial visualization 180 of the current realtime communication session.
  • the private virtual area is configured in accordance with the customization records in the interaction history.
  • the private virtual area also is populated with the other elements of the interaction history in accordance with the specification provided by the template 178.
  • each private virtual area is associated with a respective unique identifier (e.g., AreaJDI and Ar ⁇ aJD2) and is indexed by the respective identifiers (e.g., ComrnJDA, CommJDB, CommJDX, and ComrnJDY) of all the communicants who own the private virtual area,
  • a respective unique identifier e.g., AreaJDI and Ar ⁇ aJD2
  • the respective identifiers e.g., ComrnJDA, CommJDB, CommJDX, and ComrnJDY
  • each of the private virtual areas is jointly owned by a respective pair of communicants
  • Each area identifier is associated with a respective template specification identifier that uniquely identifies a particular area specification.
  • Each area identifier also is associated with a respective configuration data identifier that uniquely identifies a particular set of data (e.g., customization data) that is used by the platform to configure the private virtual area.
  • PiG. 14 shows an embodiment of a data model 182 that relates interaction records 38 in the relationship database 36 with respective ones of the private virtual areas. This relationship is used by the platform in the process of populating the private virtual area with the elements of the interaction history in accordance with the associated template specification.
  • FiGS. 15 and 16 show an embodiment of a spatia! interface 188 for realtime networked communications between communicants in a private virtual communication area (labeled "Chat with Dave") that ss created by the platform for the private bilateral interactions between the user (i.e.. Jack) and another communicant (i.e., Dave).
  • FIG. 15 depicts an exemplary state of the private virtual area in which Dave left the area after having just interacting with Jack, who still is in the private virtual area.
  • FIG. 16 depicts the state of the private virtual area in which Jack just entered the area, which already was occupied by Dave.
  • the spatial interface 188 provides a spatial visualization of the private virtual area.
  • each of the communicants is represented graphically by a respective sprite 196, 198 and the private virtual area is represented graphically by a 2.5-dimensional iconographic view of a cloud.
  • the centra! focus typically relates to matters that traditionally are associated with real-world physical spaces (e.g., work, home, meetings, clubs, etc.),
  • the private virtual area includes a
  • the graphical representation of a communicant is repositioned adjacent to the viewscreen object and a pair of eyes is added to the graphical representation to provide an additional visual indication that the associated
  • communicant is viewing an application in connection with the viewscreen object 200.
  • the communicants that are associated with the private virtual area may customize the private virtual area, for example, by adding additional props (e.g., another viewscreen prop or a table prop), changing the color scheme, etc.
  • Communicants interact with the props by selecting them with a input device (e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like), in response to a communicant's selection of a particular prop, the communicant ' s sprite either is repositioned adjacent to the selected prop or it is replicated and the replicated sprite is positioned adjacent to the selected prop and the original sprite remains where it was seated.
  • a input device e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like
  • [00981 f ⁇ e spatial interface 188 is integrated with a realtime communications interface window 190 that additionally includes a toolbar 192, a chat log area 194 a text box 206, and a Send button 208 that function in the same way as the toolbar 112, the chat log area 114, the text box 116, and the Send button 118 of the spatial interface 110 shown in FIG, 8,
  • the chat log area 194 displays a log of events that are associated with the private bilateral interactions between the user (i.e., Jack) and another one of the communicants (i.e., Dave).
  • the log of events includes sequences of text messages that the user has exchanged with the other communicant in the associated private virtual area.
  • the user may enter text messages in the text box 206 and transmit the text messages to the other communicant in the private virtual area by selecting the Send button 208.
  • An exemplary set of events that can be recorded in the chat log area 204 include; text message entries; changes in the presence status of communicants in the private virtual area; changes in the speaker and microphone settings of the
  • each of the events is associated with a respective timestamp 209 that identifies the date and time of the associated event.
  • the application sharing event description 214 has a description of the event class (Share), the identity of the sharer (Dave), the label of the share target (Screen 1 ), the URL of the share target (represented by the underlining of the share target label), the timestamp associated with the event, and a description of the shared application.
  • a graphics! separator such as ruie line 216, is added to the chat log area 194 between the events of one communication session (also referred to as a "conversation " ') and those of another communication session.
  • the textuai descriptions of prior communication sessions are
  • previous conversations are "collapsed” and labeled with the iist of participants in the conversation as weli as a timestamp of the most recent event or message within the conversation.
  • Clicking a "toggie” to the Seft of the conversation iabe! opens up the conversation and displays the fuSl contents of the conversation in the chat log area 194
  • the chat tog area 194 contains a standard "chat history" (also referred to as an "instant message history”) that includes a list of entries typed remotely by two or more networked communicants, interleaved in the order the entries have been typed.
  • the chat history typically is displayed on each communicant's terminal display, along with an indication of which user made a particular entry and at what time reiative to other communicant's entries. This provides a session history for the chat by enabling communicants to independently view the entries and the times at which each entry was made.
  • the spatial interface 188 provides a context for organizing the presentation of the events that are displayed in the chat log area 194.
  • each of the displayed events is labeled with a respective tag that visualiy correlates with the appearance of the sprite of the communicant that sourced the displayed event, in particular, each of the events that is s ⁇ urced by a particular one of the communicants is labeled with a respective icon 210, 212 with a visual appearance (e.g., color-code) that matches the vssuai appearance of that communicant's sprite, in the illustrated embodiment, for example, the color of the icon 212 matches the color of the body of Dave's sprite 198 and the color of the icon 210 matches the color of Jack's sprite 196.
  • a visual appearance e.g., color-code
  • FIG. 17 shows an embodiment of a spatial interface 220 for realtime networked communications between communicants in a private virtual area (labeled "Chat with Yuka") that is created by the platform for the private bilateral interactions between the user (i.e., Arkadi) and another communicant (i.e., Yuka).
  • the spatial interface 220 provides a spatial visualization of the private virtual area, in this
  • each of the communicants is represented graphically by a respective sprite 222, 224 and the virtual area is represented graphically by a 2.5-dimensional iconographic view of a cloud.
  • the spatial interface 220 is integrated with a realtime communications interface window 218 that additionally has the same interface elements as the interface window 190 shown in FIGS. 15 and 16, including a toolbar 192, a chat log area 194, a text box 206, and a Send button 208.
  • the private virtual area includes two viewscreen props 226, 228 and a table prop 230, on top of which is shown a graphical representation 231 of a data fife (i.e., "DE Expense Report__ml.doc") that was shared by a respective one of the communicants.
  • the communicants that are associated with the private virtual area may customize the pnvate virtual area, for example, by adding additional props (e.g., another viewscreen prop or a table prop), changing the color scheme, etc.
  • Communicants interact with the props by selecting them with an input device (e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like).
  • an input device e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like.
  • the communicant's sprite either is repositioned adjacent to the selected prop or it is replicated and the replicated sprite is positioned adjacent to the selected prop and the original sprite remains where it was seated, In the example shown in FIG.
  • FiG. 18 is a diagrammatic view of an embodiment 300 of the network communication environment 10 (see FiG. 1 ) in which the synchronous conferencing server node 30 is implemented by a virtual environment creator 302,
  • the virtual environment creator 302 includes at least one server network node 304 that provides a network infrastructure service environment 306.
  • the communications application 26 and the network infrastructure service environment 306 together provide a platform for creating a spatial virtual communication environment (also referred to herein simply as a "virtual environment" ⁇ that includes one or more of the spatia! metaphor visualizations described above.
  • the network infrastructure service environment 306 manages sessions of the first and second client nodes 12, 14 in a virtual area 308 in accordance with a virtual area application 310.
  • the virtual area application 310 is hosted by the virtual area 308 and includes a description of the virtual area 308.
  • the communications applications 26 operating on the first and second client network nodes 12, 14 present respective vtews of the virtual area 308 in accordance with data received from the network infrastructure service environment 306 an ⁇ provide respective interfaces for receiving commands from the communicants and providing a spatial interface that enhances the realtime communications between the communicants as described above.
  • the communicants typically am represented in the virtual area 308 by respective avatars, which typically move about the virtual area 308 in response to commands that are input by the communicants at their respective network nodes.
  • Each communicant's view of the virtual area 308 typically is presented from the perspective of the communicant's avatar, which increases the level of immersion experienced by the communicant.
  • Each communicant typically is able to view any part of the virtual area 308 around his or her avatar.
  • the communications applications 26 establish realtime data stream connections between the first and second client network nodes 12, 14 and other network nodes sharing the virtual area 308 based on the positions of the communicants' avatars in the virtual area 308.
  • the network infrastructure service environment 306 also maintains the relationship database 36 that contains the records 38 of interactions between communicants. Each interaction record 38 describes the context of an interaction between a pair of communicants.
  • the network 18 may include any of a local area network (LAN), a metropolitan area network (MAN) 1 and a wide area network (WAN) (e.g., the internet).
  • the network 18 typically includes a number of different computing platforms and transport facilities that support the transmission of a wide variety of different media types (e.g.. text, voice, audio, and video) between network nodes,
  • the communications application 26 typically operates on a client network node that includes software and hardware resources which, together with administrative policies, user preferences (including preferences regarding the exportation of the user's presence and the connection of the user to areas and other users), and other settings, define a local configuration that influences the administration of realtime connections with other network nodes.
  • the network connections between network nodes may be arranged in a variety of different stream handling topologies, including a peer-to-peer architecture, a server-mediated
  • the network infrastructure service environment 30 typically includes one or more network infrastructure services that cooperate with the communications applications 26 in the process of establishing and administering network connections between the client nodes 12, 14 and other network nodes (see FIGS, 1 and 18),
  • the network infrastructure services may run on a single network node or may be distributed across multiple network nodes.
  • the network infrastructure services typically run on one or more dedicated network nodes (e.g., a server computer or a network device that performs one or more edge services, such as routing and switching) In some embodiments, however, one or more of the network infrastructure services run on at least one of the communicants ' network nodes.
  • the network infrastructure services that are included in the exemplary embodiment of the network infrastructure service environment 30 are an account service, a security service, an area service, a rendezvous service, and an interaction service.
  • the account service manages communicant accounts for the virtual environment.
  • the account service also manages the creation and issuance of authentication tokens that can be used by client network nodes to authenticate themselves to any of the network infrastructure services.
  • the security service controls communicants 1 access to the assets and other resources of the virtual environment.
  • the access contro! method implemented by the security service typically is based on one or more of capabilities (where access is granted to entities having proper capabilities or permissions) and an access control list (where access is granted to entities having identities that are on the list). After a particular communicant has been granted access to a resource, that communicant typically uses the functionality provided by the other network infrastructure services to interact in the network communications environment 300.
  • the area service administers virtual areas.
  • the area service remotely configures the communications applications 26 operating on the first and second client network nodes 12, 14 in accordance with the virtual area application 308 subject to a set of constraints 312 (see FlG. 18).
  • the constraints 312 typically include controls on access to the virtual area.
  • the access controls typically are based on one or more of capabilities (where access is granted to communicants or client nodes having proper capabilities or permissions) and an access control list (where access is granted to communicants or client nodes having identities that are on the list).
  • the area service also manages network connections that are associated with the virtual area subject to the capabilities of the requesting entities, maintains global state information for the virtual area, and serves as a data server for the client network nodes participating in a shared communication session in a context defined by the virtual area 308.
  • the global state information includes a list of ail the objects that are in the virtual area and their respective locations in the virtual area.
  • the area service sends instructions that configure the client network nodes.
  • the area service also registers and transmits initialization information to other client network nodes that request to join the communication session.
  • the area service may transmit to each joining client network node a list of components (e.g., plugins) that are needed to render the virtual area 308 on the client network node in accordance with the virtual area application 310.
  • the area service also ensures that the client network nodes can synchronize to a global state if a communications fault occurs.
  • the area service typically manages communicant interactions with virtual areas via governance rules that are associated with the virtual areas.
  • the rendezvous service manages the collection, storage, and distribution of presence information and provides mechanisms for network nodes to communicate with one another (e.g., by managing the distribution of connection handles) subject to the capabilities of the requesting entities.
  • the rendezvous service typically stores the presence information in a presence database.
  • the rendezvous service typically manages communicant interactions with each other via communicant privacy preferences.
  • the interaction service maintains the relationship database 36 that contains the records 38 of interactions between communicants. For every interaction between communicants, one or more services of the network infrastructure service environment 306 (e.g., the area service) transmit interaction data to the interaction service. In response, the interaction service generates one or more respective interaction records and stores them in the relationship database. Each interaction record describes the context of an interaction between a pair of communicants.
  • an interaction record contains an identifier for each of the communicants, an identifier for the place of interaction (e.g., a virtual area instance), a description of the hierarchy of the interaction place (e.g., a description of how the interaction room relates to a larger area), start and end times of the interaction, and a list of all files and other data streams that are shared or recorded during the interaction.
  • the interaction service tracks when it occurred, where it occurred, and what happened during the interaction in terms of communicants involved ⁇ e.g., entering and exiling), objects that are activated/deactivated, and the files that were shared.
  • the interaction service also supports queries on the relationship database 36 subject to the capabilities of the requesting entities.
  • the interaction service presents the results of queries on the interaction database records in a sorted order (e.g.. most frequent or most recent) based on virtual area.
  • the query results can be used to drive a frequency sort of contacts whom a communicant has met in which virtual areas , as well as sorts of who the communicant has met with regardless of virtual area and sorts of the virtual areas the communicant frequents most often.
  • the query results also may be used by application developers as part of a heuristic system that automates certain tasks based on relationships.
  • An example of a heuristic of this type is a heuristic that permits communicants who have visited a particular virtual area more than five times to enter without knocking by default, or a heuristic that allows
  • Queries on the relationship database 36 can be combined with other searches. For example, queries on the relationship database may be combined with queries on contact history data generated for interactions with contacts using a communication system (e.g., Skype, Facebook, and Flickr) that is outside the domain of the network infrastructure service environment 306.
  • a communication system e.g., Skype, Facebook, and Flickr
  • the virtual area instance may correspond to an abstract (non-geometric) virtual space that is defined with respect to abstract coordinates.
  • the virtual area instance may correspond to a visual virtual space that is defined with respect to one-, two- or three- dimensional geometric coordinates that are associated with a particular visualization.
  • Abstract virtual areas may or may not be associated with respective visualizations, whereas visual virtual areas are associated with respective visualizations.
  • communicants typically are represented by respective avatars (e.g., sprites) in a virtual area that has an associated visualization
  • the avatars move about the virtual area in response to commands that are input by the communicants at their respective network nodes.
  • the communicant's view of a virtual area instance typically is presented from the
  • each communicant typically is able to view any part of the visual virtual area around his or her avatar, increasing the level of immersion that is experienced by the communicant
  • a virtual area typicaily includes one or more zones that are associated with respective rules that govern the switching of realtime data streams between the network nodes that are represented by the avatars in the virtual area.
  • the switching rules dictate how local connection processes executing on each of the network nodes establishes communications with the other network nodes based on the locations of the communicants 3 avatars in the zones of the virtual area.
  • a virtua! area typically is defined by a specification that includes a description of geometric elements of the virtual area and one or more rules, including switching rules and governance rules.
  • the switching rules govern realtime stream connections between the network nodes.
  • the governance ruies controi a communicant's access to resources, such as the virtual area itself, regions with the virtual area ; and objects within the virtual area.
  • the geometric elements of the virtua! area are described in accordance with the COLLADA - Digital Asset Schema Release 1.4.1 April 2006 specification ⁇ availabie from http://www.khronos.org/coilada/), and the switching ruies are described using an extensible markup language (XML) text format (referred to herein as a virtual space description format (VSDL)) in accordance with the COLLADA Streams Reference specification described in US Application Nos. 11/923,629 and 11/923 ; 634.
  • XML extensible markup language
  • VSDL virtual space description format
  • [001231 ⁇ e geometric eiements of the virtua! area typically include physical geometry and collision geometry of the virtual area.
  • the physical geometry describes the shape of the virtual area.
  • the physical geometry typically is formed from surfaces of triangles, quadrilaterals, or polygons. Colors and textures are mapped onto the physical geometry to create a more realistic appearance for the virtua! area. Lighting effects may be provided, for example, by painting Sights onto the visual geometry and modifying the texture, color, or intensity near the lights.
  • the coilision geometry describes invisible surfaces that determine the ways in which objects can move in the virtual area.
  • the collision geometry may coincide with the visual geometry, correspond to a simpler approximation of the visuai geometry, or relate to application-specific requirements of a virtual area designer.
  • the switching rules typically include a description of conditions for connecting sources and sinks of realtime data streams in terms of positions in the virtual area.
  • Each rule typically includes attributes that define the realtime data stream type to which the rule applies and the iocation or locations in the virtual area where the rule applies.
  • each of the rules optionally may include one or more attributes that specify a required role of the source, a required role of the sink, a priority level of the stream, and a requested stream handling topology.
  • one or more implicit or default switching rules may apply to that part of the virtual area.
  • One exemplary default switching rule is a rule that connects every source to every
  • Policy ruies may apply globally to all connections between the client nodes or only to respective connections with individual client nodes.
  • An example of a policy rule is a proximity policy rule that only allows connections of sources with compatible sinks that are associated with respective objects that are within a prescribed distance (or radius) of each other in the virtual area.
  • governance rules are associated with a virtual area to control who has access to the virtual area, who has access to its contents, what is the scope of that access to the contents of the virtual area (e.g., what can a user do with the contents), and what are the follow-on consequences of accessing those contents (e.g., record keeping, such as audit logs, and payment requirements),
  • an entire virtual area or a zone of the virtual area is associated with a "governance mesh.”
  • a governance mesh is implemented in a way that is analogous to the implementation of the zone mesh described in US
  • a governance mesh enables a software application developer to associate governance rules with a virtual area or a zone of a virtual area. This avoids the me ⁇ for the creation of individual permissions for every file in a virtual area and avoids the need to deal with the complexity that potentially could arise when there is a need to treat the same document differently depending on the context.
  • a virtual area is associated with a governance mesh that associates one or more zones of the virtual area with a digital rights management (DRM) function.
  • the DRIVI function controls access to one or more of the virtual area or one or more zones within the virtual area or objects within the virtual area.
  • the DRM function is triggered every time a communicant crosses a governance mesh boundary within the virtual area. The DRM function determines whether the triggering action is permitted and, if so, what is the scope of the permitted action, whether payment is needed, and whether audit records need to be generated.
  • the associated governance mesh is configured such that if a communicant is able to enter the virtual area he or she is able to perform actions on all the documents that are associated with the virtual area, including manipulating the documents, viewing the documents, downloading the documents, deleting the documents, modifying the documents and re-upioading the documents.
  • the virtual area can become a repository for information that was shared and discussed in the context defined by the virtual area
  • the communications application 26 includes: a, local Human Interface Devices (HiDs) and audio playback devices: b, a So3D graphical display, avatar, and physics engine; c, a system database and storage facility.
  • HiDs Human Interface Devices
  • audio playback devices b, a So3D graphical display, avatar, and physics engine
  • c a system database and storage facility.
  • the local HIDs enable a communicant to input commands and other signals into the client network node while participating in a virtual area communications session.
  • Exemplary HiDs include a computer keyboard, a computer mouse, a touch screen display, and a microphone
  • the audio playback devices enable a communicant to playback audio signals that are received during a virtual area communications session.
  • Exemplary audio playback devices include audio processing hardware (e.g., a sound card) for manipulating (e.g., mixing and applying special effects) audio signals, and speakers for outputting sounds.
  • audio processing hardware e.g., a sound card
  • speakers for outputting sounds.
  • the So3D engine is a three-dimensional visualization engine that controls the presentation of a respective view of a virtual area and objects in the virtual area on a display monitor.
  • the So3D engine typically interfaces with a graphical user interface driver and the HID devices to present the views of the virtual area and to allow the communicant to control the operation of the communications application 26.
  • the So3D engine receives graphics rendering instructions from the area service.
  • the So3D engine also may read a local
  • the So3D engine Based on this information, the So3D engine generates a visual representation (i.e., an image) of the virtual area and the objects in the virtual area from the point of view (position and orientation) of the communicant's avatar in the virtual area.
  • the visual representation typically is passed to the graphics rendering components of the operating system, which drive the graphics rendering hardware to render the visual representation of the virtual area on the client network node.
  • ⁇ ⁇ e communicant can control the presented view of the virtual area by inputting view control commands via a HID device (e.g., a computer mouse)
  • the So3D engine updates the view of the virtual area in accordance with the view control commands.
  • the So3D engine also updates the graphic representation of the virtual area on the display monitor in accordance with updated object position information received from the area service.
  • the system database and storage facility stores various kinds of information that is used by the platform.
  • Exemplary information that typically is stored by the storage facility includes the presence database, the relationship database, an avatar database, a real user id (RUID) database, an art cache database, and an area application database. This information may be stored on a single network node or it may be distributed across multiple network nodes.
  • a communicant typically connects to the network 18 from a client network node.
  • the client network node typicaily is implemented by a general-purpose computer system or a dedicated communications computer system (or ''console", such as a network-enabled video game console).
  • the client network node executes communications processes that establish realtime data stream connections with other network nodes and typically executes visualization rendering processes that present a view of each virtual area entered by the communicant.
  • FIG. 19 shows an embodiment of a client network node that is implemented by a computer system 320.
  • the computer system 320 includes a processing unit 322 : a system memory 324, and a system bus 326 that couples the processing unit 322 to the various components of the computer system 320.
  • the processing unit 322 may include one or more data processors, each of which may be in the form of any one of various commercially available computer processors.
  • the system memory 324 includes one or more computer-readable media that typically are associated with a software application addressing space that defines the addresses that are available to software applications.
  • the system memory 324 may include a read only memory (ROM) that stores a basic input/output system (BIOS) that contains startup routines for the computer system 320, and a random access memory (RAM).
  • the system bus 326 may be a memory bus, a peripheral bus or a local bus, and may be compatible with any of a variety of bus protocols, including PCi, VESA, MicroChannel. ISA, and EISA.
  • the computer system 320 also includes a persistent storage memory 328 ⁇ e.g.
  • a communicant may interact (e.g., input commands or data) with the computer system 320 using one or more input devices 330 (e.g. one or more
  • GUI graphical user interface
  • the computer system 320 also may include other input/output hardware (e.g., peripheral output devices, such as speakers and a printer).
  • the computer system 320 connects to other network nodes through a network adapter 336 (also referred to as a "network interface card” or NSC).
  • NSC network interface card
  • a number of program modules may be stored in the system memory 324, including application programming interfaces 338 (APis), an operating system (OS) 340 (e.g., the Windows XP ⁇ operating system available from Microsoft Corporation of Redmond, Washington U.S.A.), the communications application 2S 1 drivers 342 (e.g., a GUi driver), network transport protocols 344, and data 348 (e.g., input data, output data, program data, a registry, and configuration settings).
  • API application programming interfaces 338
  • OS operating system
  • the communications application 2S 1 drivers 342 e.g., a GUi driver
  • network transport protocols 344 e.g., input data, output data, program data, a registry, and configuration settings.
  • the one or more server network nodes of the virtual environment creator 16 are implemented by respective general-purpose computer systems of the same type as the client network node 120, except that each server network node typically includes one or more server software applications.
  • the one or more server network nodes of the virtual environment creator 16 are implemented by respective network devices that perform edge services (e.g., routing and switching).
  • edge services e.g., routing and switching
  • each of the client network nodes generates a respective set of realtime data streams (e.g., motion data streams, audio data streams, chat data streams, file transfer data streams, and video data streams).
  • each communicant manipulates one or more input devices (e.g., the computer mouse 52 and the keyboard 54) that generate motion data streams, which control the movement of his or her avatar in the virtual area 66, Sn addition, the communicant's voice and other sounds that are generated locally in the vicinity of the computer system 48 are captured by the microphone 60.
  • the microphone 60 generates audio signals that are converted into realtime audio streams.
  • Respective copies of the audio streams are transmitted to the other network nodes that are represented by avatars in the virtual area 66.
  • the sounds that are generated locally at these other network nodes are converted into realtime audio signals and transmitted to the computer system 48.
  • the computer system 48 converts the audio streams generated by the other network nodes into audio signals that are rendered by the speakers 56, 58.
  • the motion data streams and audio streams may be transmitted from each of the communicant nodes to the other client network nodes either directly or indirectly.
  • each of the client network nodes receives copies of the realtime data streams that are transmitted by the other client network nodes, in other stream handling topologies, one or more of the client network nodes receives one or more stream mixes that are derived from realtime data streams that are sourced (or originated) from other ones of the network nodes.
  • the area service maintains global state information that includes a current specification of the virtual area, a current register of the objects that are in the virtual area, and a list of any stream mixes that currently are being generated by the network node hosting the area service.
  • the objects register typically includes for each object in the virtual area a respective object identifier (e.g., a label that uniquely identifies the object), a connection handle (e.g., a URI, such as an IP address) that enables a network connection to be established with a network node that is associated with the object, and interface data that identifies the realtime data sources and sinks that are associated with the object (e.g., the sources and sinks of the network node that is associated with the object).
  • a respective object identifier e.g., a label that uniquely identifies the object
  • a connection handle e.g., a URI, such as an IP address
  • the objects register aiso typically includes one or more optional role identifiers for each object; the role identifiers may be assigned explicitly to the objects by either the communicants or the area service, or may be inferred from other attributes of the objects or the user.
  • the objects register also includes the current position of each of the objects in the virtual area as determined by the area service from an analysis of the realtime motion data streams received from the network nodes associated with objects in the virtual area Sn this regard, the area service receives realtime motion data streams from the network nodes associated with objects in the virtual area, tracks the communicants' avatars and other objects that enter, leave, and move around in the virtual area based on the motion data. The area service updates the objects register in accordance with the current locations of the tracked objects.
  • the area service maintains for each of the client network nodes a set of configuration data : including interface data, a zone list, and the positions of the objects that currently are in the virtual area.
  • the interface data includes for each object associated with each of the client network nodes a respective list of all the sources and sinks of realtime data stream types that are associated with the object
  • the zone list is a register of all the zones in the virtual area that currently are occupied by the avatar associated with the corresponding client network node.
  • the communications application 26 also includes a graphical navigation and interaction interface (referred to herein as a "seeker interface") that interfaces the user with the spatial virtual communication environment.
  • the seeker interface includes navigation controls that enable the user to navigate the virtual environment and interaction controls that enable the user to control his or her interactions with other communicants in the virtual communication environment.
  • the navigation and interaction controls typically are responsive to user selections that are made using any type of input device, including a computer mouse, a touch pad, a touch screen display, a keyboard, and a video game controller
  • the seeker interface is an application that operates on each client network node.
  • the seeker interface is a small, lightweight interface that a user can keep up and running aii the time on his or her desktop.
  • the seeker interface allows the user to launch vtrtuaS area applications and provides the user with immediate access to realtime contacts and realtime collaborative places (or areas)
  • the seeker interface is integrated with realtime communications applications and/or realtime communications components of the underlying operating system such that the seeker interface can initiate and receive realtime communications with other network nodes.
  • a virtual area is integrated with the user's desktop through the seeker interface such that the user can upload files into the virtual environment created by the virtual environment creator 16, use files stored in association with the virtual area using the native client software applications independently of the virtual environment while still present in a virtual area, and more generally treat presence and position within a virtual area as an aspect of their operating environment analogous to other operating system functions rather than just one of several applications.
  • any of the embodiments of the spatial interfaces that are described herein may be integrated into the seeker interface in order to provide a context for depicting the current communication of the communicants involved in realtime networked communications.
  • Embodiments of these spatial interfaces also provide a context for organizing the presentation of various interface elements that are used by communicants to participate in realtime networked communications, as described above.
  • the embodiments that are described herein provide improved systems and methods for visualizing rea ⁇ time network communications, in particular, these embodiments apply a spatial metaphor on top of realtime networked communications.
  • the spatial metaphor provides a context for depicting the current communication state of the communicants involved in realtime networked communications.
  • the spatial metaphor also provides a context for organizing the presentation of various interface elements that are used by communicants to participate in realtime networked communications.

Abstract

A current realtime communication session is established between communicants operating on respective network nodes (12, 14). A spatial visualization (70, 92, 188, 220) of the current realtime communication session is displayed. The spatial visualization (70, 92, 188, 220) includes a graphical representation (46, 48) of each of the communicants in spatial relation to a graphical representation (50) of a virtual area. During the current communication session, visual cues are depicted in the spatial visualization (70, 92, 188, 220) that show current communication states of the communicants, where each of the communication states corresponds to a state of a respective communication channel over which a respective one of the communicants is configured to communicate.

Description

SPATIAL INTERFACES FOR REALTIME NETWORKED
CGWϊMUNJCATIONS
BACKGROUND
[00011 When face-to-face communications are not practical, people often rely on one or more technological solutions to meet their communications needs. These solutions typically are designed to simulate one or more aspects of face-to-face communications. Traditionai telephony systems enable voice communications between callers, instant messaging (also referred to as "chat") communications systems enable users to communicate text messages in real time through instant message computer clients that are interconnected by an instant message server. Some instant messaging systems additionally allow users to be represented in a virtual environment by user- controllable graphic objects (referred to as ''avatars"). Interactive virtual reality communication systems enable users in remote locations to communicate over multiple real-time channels and to interact with each other by manipulating their respective avatars in three-dimensional virtual spaces. What are needed are improved interfaces for realtime network communications
SUWiWARY
[0002J in one aspect, the invention features a method in accordance with which a current realtime communication session is established between communicants operating on respective network nodes A spatial visualization of the current realtime communication session is displayed. The spatial visualization includes a graphical representation of each of the communicants in spatial relation to a graphical
representation of a virtual area. During the current communication session, visual cues are depicted in the spatial visualization that show current communication states of the communicants, where each of the communication states corresponds to a state of a respective communication channel over which a respective one of the communicants is configured to communicate.
[QQQ3J in another aspect, the invention features a method in accordance with which a current realtime communication session is established between communicants operating on respective network nodes. A spatial visualization of the current realtime communication session is displayed. The spatial visualization includes a graphical representation of each of the communicants in spatiai relation to a graphical representation of a virtual area. During the current communication session, a log of event descriptions is presented. The event descriptions describe respective events involving interactions of the communicants in the virtual area. The event descriptions are presented in contextua! association with elements of the spatial visualization of the current realtime communication session.
[0004] In another aspect, the invention features a method in accordance with which receipt of a command from a first communicant operating on a first network node to initiate a private communication with a second communicant operating on a second network node prompts a response that includes the following, A current realtime communication session is established between the first and second network nodes, A private virtual area associated with the first and second communicants is identified. Context configuration data associated with the private virtual area and generated in response to interactions of the first and second communicants in the private virtual area is retrieved, A spatial visualization of the current realtime communication session is displayed The spatial visualization includes graphical representations of the first and second communicants in spatial relation to a graphical representation of the virtual area configured in accordance with the context configuration data.
[0005J The invention also features apparatus operabie to implement the method described above and computer-readable media storing computer-readable instructions causing a computer to implement the method described above,
DESCRtPTtON OF DRAWINGS
[0006] FiG 1 is a diagrammatic view of an embodiment of a network communication environment that includes a first client network node, a second client network node, and a synchronous conferencing server node.
[0007] FiG. 2 is a flow diagram of an embodiment of a method of visualizing realtime networked communications on a client network node.
[0008] FiGS. 3A-3D. 4. and 5 are diagrammatic views of spatial interfaces for realtime networked communications,
[0009] FiG. 6 is a diagrammatic view of an embodiment of a spatial interface for realtime networked communications.
[0010] FiG. 7 is a flow diagram of an embodiment of a method of managing realtime networked communications. [0011 ] FiG. 8 is a diagrammatic view of an embodiment of a spatial interface integrated with a realtime communications interface.
[0012| FIG. 9 is a diagrammatic view of an embodiment of the spatial interface shown in FlG, 8 integrated with an additional spatial interface,
[00131 FiG. 10 is a diagrammatic view of an embodiment of a graphical user interface.
[00143 FIG. 11 is a flow diagram of an embodiment of a method of managing realtime networked communications between networked communicants in a private virtual area.
[00151 FiG. 12 is a diagrammatic view of an embodiment of a process of generating a spatial visualization of a current realtime communication session.
[00163 PiG. 13 is a diagrammatic view of an embodiment of a data model relating area identifiers to communicants, template specifications, and context data.
[00171 FiG. 14 is a diagrammatic view of an embodiment of a data model relating interaction record identifiers with area identifiers and interaction records.
[0018| FlG. 15 is a diagrammatic view of an embodiment of a spatial interface integrated with a realtime communications interface for realtime networked
communications in a private virtual area.
[00191 FiG. 16 is a diagrammatic view of an embodiment of the spatial interface shown in FiG. 15.
[00203 FIG. 17 is a diagrammatic view of an embodiment of a spatial interface integrated with a realtime communications interface for realtime networked
communications in a private virtual area.
[00211 FiG 18 is a diagrammatic view of an embodiment of a network communication environment that includes a first client network node, a second client network node, and a virtual environment creator.
[00221 FiG. 19 is a block diagram of the network communication environment of FIG, 1 that shows components of an embodiment of a client network node,
DETAILED DESCRIPTION
[00231 tπ the following description, like reference numbers are used to identify like elements. Furthermore, the drawings are intended to illustrate major features of exemplary embodiments in a diagrammatic manner. The drawings are not intended to depict every feature of actual embodiments nor relative dimensions of the depicted elements, and are not drawn to scale.
i. DEFINIITON OF TERMS
[0024] A "communicant" is a person who communicates or otherwise interacts with other persons over one or more network connections, where the communication or interaction may or may not occur in the context of a virtual area. A "user" is a
communicant who is operating a particular network node that defines a particular perspective for descriptive purposes.
[00251 A 'realtime contact" of a user is a communicant or other person who has communicated with the user via a realtime communications platform.
[0026] A ^computer" is any machine, device, or apparatus that processes data according to computer-readable instructions that are stored on a computer-readable medium either temporariiy or permanently, A "computer operating system" is a software component of a computer system that manages and coordinates the performance of tasks and the sharing of computing and hardware resources, A "software application" (also referred to as software, an application, computer software, a computer application, a program, and a computer program} is a set of instructions that a computer can interpret and execute to perform one or more specific tasks. A "computer data fiie11 is a block of information that durably stores data for use by a software application.
[0027] A "window" is a visual area of a display that typically includes a user interface. A window typically displays the output of a software process and typically enables a user to input commands or data for the software process. A window that has a parent is called a "child window." A window that has no parent, or whose parent is the desktop window, is calied a "top-level window." A "desktop" is a system-defined window that paints the background of a graphical user interface (GUS) and serves as the base for all windows displayed by all software processes.
[00281 A "database'5 is an organized collection of records that are presented in a standardized format that can be searched by computers. A database may be stored on a single computer-readable data storage medium on a single computer or it may be distributed across multiple computer-readable data storage media on one or more computers.
[00291 A "data sink" (referred to herein simply as a "sink") is any of a device (e.g., a computer), part of a device, or software that receives data. [0030] A "data source5' (referred to herein simply as a "source") is any of a device (e.g., a computer), part of a device, or software that originates data.
[00313 A ^network node" (also referred to simpiy as a "node") is a junction or connection potnt in a communications network. Exemplary network nodes include, but are not limited to, a terminal a computer, and a network switch, A "server" network node is a host computer on a network that responds to requests for information or service. A "client" network node is a computer on a network that requests information or service from a server. A "network connection" is a link between two communicating network nodes. The term local network node" refers to a network node that currently is the primary subject of discussion. The term "remote network node" refers to a network node that is connected Io a local network node by a network communications link, A "connection handle" is a pointer or identifier {e.g., a uniform resource identifier (URi)) that can be used to establish a network connection with a communicant, resource, or service on a network node. A "network communication" can include any type of information (e g , text, voice: audio, video, electronic mail message, data filβ; motion data stream, and data packet) that is transmitted or otherwise conveyed from one network node to another network node over a network connection.
[0032J Synchronous conferencing refers to communications in which communicants participate at the same time. Synchronous conferencing encompasses all types of networked collaboration technologies, including instant messaging (e.g., text chat), audio conferencing, video conferencing, application sharing, and file sharing technologies.
[0033] A "communicant interaction8 is any type of direct or indirect action or influence between a communicant and another network entity, which may include for example another communicant, a virtual area, or a network service. Exemplary types of communicant communications include communicants communicating with each other in realtime, a communicant entering a virtual area, and a communicant requesting access to a resource from a network service.
[0034] "Presence" refers to the ability and willingness of a networked entity (e.g., a communicant, service, or device) to communicate, where such willingness affects the ability to detect and obtain information about the state of the entity on a network and the ability to connect to the entity. [0035] A "realtime data stream" is data that is structured and processed in a continuous flow and is designed to be received with no delay or oniy imperceptible delay. Reaitime data streams include digital representations of voice, video, user movements, facial expressions and other physicai phenomena, as well as data within the computing environment that may benefit from rapid transmission, rapid execution, or both rapid transmission and rapid execution, including for example, avatar movement instructions, text chat, realtime data feeds (e.g., sensor data, machine control instructions, transaction streams and stock quote information feeds), and file transfers,
[0036] A "link" is a connection between two network nodes and represents the fυli bandwidth aliocated by the two nodes for real-time communication. Each link is divided into channeis that carry respective real-time data streams. Channels are aliocated to particuiar streams within the overall bandwidth that has been allocated to the link.
[0037] A "virtuai area" (aiso referred to as an "area" or a <kplace") is a representation of a computer-managed space or scene. Virtuai areas typically are one- dimensional, two-dimensional, or three-dimensional representations , although in some embodiments a virtual area may correspond to a single point. Oftentimes, a virtua! area is designed to simuiate a physical, real-world space. For example, using a traditional computer monitor, a virtual area may be visualized as a two-dimensional graphic of a three-dimensional computer-generated space. However, virtuai areas do not require an associated visualization to impiement switching rules. A virtual area typically refers to an instance of a virtual area schema, where the schema defines the structure and contents of a virtuai area in terms of variables and the instance defines the structure and contents of a virtual area in terms of values that have been resolved from a particular context,
[0038] A "virtual area application" (also referred to as a "virtual area
specification") is a description of a virtual area that is used in creating a virtual environment. The virtual area application typically includes definitions of geometry, physics, and realtime switching ruies that are associated with one or more zones of the virtual area.
[0039] A "virtuai environment" is a representation of a computer-managed space that includes at least one virtual area and supports realtime communications between communicants. [0040] A "zone" is a region of a virtual area that is associated with at least one switching rule or governance rule. A "switching rule" is an instruction that specifies a connection or disconnection of one or more realtime data sources and one or more realtime data sinks subject to one or more conditions precedent, A switching rule controls switching (e.g., routing, connecting, and disconnecting) of realtime data streams between network nodes communicating tn the context of a virtual area. A governance rule controls a communicant's access to a resource (e.g., an area, a region of an area, or the contents of that area or region), the scope of that access, and follow- on consequences of that access (e.g., a requirement that audit records relating to that access must be recorded). A "renderable zone" is a zone that is associated with a respective visualization,
[00413 A "position" in a virtual area refers to a location of a point or an area or a volume in the virtual area. A point typically is represented by a single set of one- dimensional, two-dimensional, or three-dimensional coordinates (e.g., x, y, z) that define a spot in the virtual area. An area typically is represented by the three-dimensional coordinates of three or more coplanar vertices that define a boundary of a closed two- dimensional shape in the virtual area, A volume typically is represented by the three- dimensional coordinates of four or more non-copianar vertices that define a closed boundary of a three-dimensional shape in the virtual area,
[00423 A "spatial state" is an attribute that describes where a user has presence in a virtual area. The spatial state attribute typicaϋy has a respective vaiue (e.g., a zoneJD vaiue) for each of the zones in which the user has presence.
[0043] A "communication state3 is an attribute that describes a state of a respective communication channel over which a respective one of the communicants is configured to communicate.
[0044] Sn the context of a virtual area, an "object" (also sometimes referred to as a "prop") is any type of discrete element in a virtual area that may be usefully treated separately from the geometry of the virtual area. Exemplary objects include doors, portals, windows, view screens, and speakerphone. An object typically has attributes or properties that are separate and distinct from the attributes and properties of the virtual area. An "avatar" is an object that represents a communicant in a virtual area. [0045] As used herein, the term "includes" means includes but not limited to, the term "including" means including but not limited to. The term "based on" means based at least in part on.
II. OVERVIEW
A. INTRODUCTiON
[0046J The embodiments that are described herein provide improved systems and methods for visualizing reaitime network communications, in particular, these embodiments apply a spatial metaphor on top of reaitime networked communications. The spatial metaphor provides a context for depicting the current communication states of the communicants involved in reaitime networked communications. The spatial metaphor aiso provides a context for organizing the presentation of various interface elements that are used by communicants to participate in realtime networked
communications.
[0047] FIG. 1 shows an embodiment of an exemplary network
communications environment 10 that inciudes a first client network node 12 (Client Node A), a second ciient network node 14 (Client Network Node B), and a synchronous conferencing server 16 that are interconnected by a network 18. The first client network node 12 includes a computer-readable memory 20; a processor 22, and input/output (I/O) hardware 24 (including a display). The processor 22 executes at least one communications appiication 26 that is stored in the memory 20. The second client network node 14 typically is configured in substantially the same way as the first client network node 12. in some embodiments, the synchronous conferencing server 18 manages realtime communication sessions between the first and second client nodes 12, 14. The network infrastructure service environment 30 aiso maintains a relationship database 36 that contains records 38 of interactions between communicants. Each interaction record 38 describes the context of an interaction between a pair of communicants. As explained in detail below, the communications application 26 and the synchronous conferencing server 16 together provide a platform (referred to herein as "the platform") for creating a spatial visualization context that enhances realtime communications between communicants operating on the network nodes 12, 14.
[0048] FIG. 2 shows an embodiment of a method that is implemented by the communications application 26 operating on one or both of the first and second network nodes 12, 14. This process typically is performed in response to a request from a communicant on one of the network nodes 12, 14 to initiate a realtime communication session with another communicant operating on the other network node. The
communications appiication 26 establishes a current realtime communication session between communicants operating on respective network nodes (FIG. 2, block 40) On a display, the communications application 26 dispiays a spatial visualization of the current realtime communication session (FSG. 2, biock 40). The spatia! visualization includes a graphical representation of each of the communicants in spatial relation to a graphical representation of a virtual area. The virtuai area may be represented graphically by any type of one-dimensional, two-dimensional, or three-dimensional view that situates the graphicai representations of the communicants in respective positions in a visual space. During the current communication session, the communications application 26 depicts visual cues in the spatiai visualization that show current communication states of the communicants (FlG, 2S block 44). Each of the communication states typically
corresponds to a state of a respective communication channel (e.g., text chat, audio, video, appiication share, and file share channel) over which a respective one of the communicants is configured to communicate.
[00493 in some embodiments, a log of event descriptions that describe respective events involving interactions of the communicants in the virtual area is presented on the display in contextual association with elements of the spatial visualization of the current realtime communication session. The log of event descriptions and the graphicai representation of the virtual area typically are displayed in a single graphicai user interface window. The iog of event descriptions may include, for example, at least one of: text of a chat conversation between the communicants in the virtual area; a description of a data file shared by a respective one of the
communicants in the virtual area; and a description of an application shared by a respective one of the communicants in the virtuai area. The event descriptions in the log typically are visually associated with respective ones of the graphical
representations of the communicants involved in the events described by the respective event descriptions. For example, in some embodiments, a respective label is
associated with each of the event descriptions, where the respective label has a respective visual appearance that matches a visual element of the graphical
representation of the communicant involved in the event described by the respective event description. The log of event descriptions typically are stored in one or more database records that are indexed by an identifier of the virtual area
[0050] tn some embodiments, one or more props are displayed in the virtual area, where each prop represents a respective communication channel for realtime communications between the communicants during the communication session. For example, a communicant-seiectabie table prop may be displayed in the virtual area, and a fiie share session between the communicants may be initiated in response to selection of the table prop by one of the communicants; or a communicant-selectable viewscreen prop may be displayed in the virtual area, and an application sharing session may be initiated between the communicants in response to selection of the viewscreen prop by one of the communicants. In some embodiments, a spatial property of the graphicai representation of a respective one of the communicants in relation to a respective one of the props is changed in response to selection of the respective prop by the respective communicant. For example, the graphical representation of the respective communicant may be depicted adjacent the selected prop, it may be reoriented to face the selected prop, and/or the graphical representation of the communicant may be changed (e.g., a pair of eyes may be added to the body of a communicant's sprite when it is positioned adjacent to a viewscreen prop, as shown in FIGS. 15 and 16).
[0051 J In some embodiments, a realtime instant messaging communication channel is established between the communicants during the current communication session, in these embodiments, a current chat log of a current chat conversation between the communicants occurring during the current communication session typically is displayed in association with the graphical representation of the virtual area A respective prior chat log of a prior chat conversation that occurred during a prior communication session between the communicants in the virtual area typically is displayed in association with the current chat log. The graphical representation of a given one of the communicants may be dynamically modulated in response to receipt of a respective realtime chat stream from the given communicant over the realtime instant messaging communication channel such that the current communication state of the given communicant is reflected in the dynamic modulation of the graphical
representation of the given communicant. [0052] in some embodiments, a graphical representation of a file sharing prop is displayed in the virtual area. In response to selection of the file sharing prop by a respective one of the communicants, the graphical representation of the respective communtcant typically is depicted adjacent the file sharing prop and a realtime file sharing session typically is initiated in the virtual area, A data file shared by the respective communicant during the realtime file sharing session typically is stored in a data storage device with an index that includes an identifier of the virtual area, and a communicant-selectable graphical representation of the data file typically is displayed on the file sharing prop. A download of the data file to the network node from which a given one of the communicants is operating typically is initiated in response to selection of the graphical representation of the file by the given communicant,
[0053| In some embodiments, a graphical representation of an application sharing prop is displayed in the virtual area. In response to selection of the application sharing prop by a respective one of the communicants, the graphical representation of the respective communicant typically is depicted adjacent the application sharing prop and a realtime application sharing session typically is initiated in the virtual area.
Screen shots from the network node from which the respective communicant is operating typically are shared with one or more of the other communicants during the realtime application sharing session, A graphical indication that an application that is being shared typically is displayed in connection with the application sharing prop, in some embodiments, a first graphical representation of the application sharing prop is displayed during periods of application sharing between the communicants in the virtual area and a second graphical representation of the application sharing prop different from the first graphical representation is displayed during periods free of application sharing between the communicants.
[0054] In some embodiments, in response to a command from a given one of the communicants to activate an audio sink communication channel, a realtime audio communication channel is established between the given communicant and one or more of the other communicants configured as audio sources, and the depicting graphical representation of the given communicant is modified to show that the given communicant is configured as an audio sink. Analogously, in response to a command from a given one of the communicants to activate an audio source communication channel, a realtime audio communication channel is established between the given communicant and one or more of the other communicants configured as audio sinks, and the graphical representation of the given communicant is modified to show that the given communicant is configured as an audio source.
[0055J in some embodiments, a static view of the graphica! representation of the virtual area is displayed throughout the current communication session, and the communicants are unable to navigate the graphical representations of the
communicants outside the static view of the virtual area,
[0056] in some embodiments, in response to receipt of a command from a first one of the communicants to initiate a private communication with a second one of the communicants, the current realtime communication session between the first and second communicants is established, and the graphical representations of the first and second communicants are dispiayed in spatial relation to a graphica! representation of a virtual area that is indexed by identifiers of the first and second communicants.
[0057J In some embodiments, an end state of a prior realtime communication session between the communicants is determined from data that is indexed by an identifier of the virtual area and that describes events that occurred during a prior communication session between the communicants, and the graphical representation of a virtual area is displayed in a state that corresponds to the determined end state of the prior communication session between the communicants.
B. EXEMPLARY SPATIAL INTERFACES FOR REALTIME
COMMUNICATION SESSIONS
[00581 FIGS. 3A-3D respectively show embodiments of spatial visualizations of a realtime communication session that include visual cues that reveal the current communication states of two networked communicants involved in the realtime communication session. In these embodiments, the spatial visualizations include a graphica! representation 46, 48 of each of the communicants in spatial relation to a graphical representation 50 of a virtual area. In particular, the virtual area is
represented by a perspective view of a three-dimensional visual space in which the graphical representations 46, 48 of the communicants can have different respective positions, in the illustrated embodiments, each communicant is represented by a respective circular sprite 48, 48. The states of various communication channels over which the respective communicant is configured to communicate are revealed by visual cues that are shown in the spatial visualization. For example, the on or off state of a communicant's !oca! speaker channel is depicted by the presence or absence of a headphones graphic 52 on the communicant's sprite 46. Thus, when the speakers of the communicant who is represented by the sprite 46 are on, the headphones graphic 52 is present (as shown in FIG. 3B) and, when the communicant's speakers are off, the headphones graphic 52 is absent (as shown in FiG. 3A). The on or off state of the communicant's microphone is depicted by the presence or absence of a microphone graphic 54 on the communicant's sprite 46 and a series of concentric circles 56 that radiate away from the communicant's sprite 46 in a series of expanding waves. Thus, when the microphone is on, the microphone graphic 54 and the radiating concentric circles 58 are present (as shown in FIG. 3C) and, when the microphone is off, the microphone graphic 54 and the radiating concentric circles 56 are absent (as shown in FIGS. 3A, 3B, and 3D). The headphones graphic 52, the microphone graphic 54, and the radiating concentric circles 56 serve as visuai cues of the states of the
communicant's sound playback and microphone devices. The on or off state of a communicant's text chat channel is depicted by the presence or absence of a hand graphic 57 adjacent the communicant's sprite (as shown in FIG. 3D). When a communicant is transmitting text chat data to another network node the hand graphic 57 is present, and when a communicant is not transmitting text chat data the hand graphic 57 is not present. Sn some embodiments, text chat data is transmitted only when keyboard keys are depressed, in which case the visualization of the communicant s text channel appears as a flashing on and off of the hand graphic 57.
[0059] FIGS. 4 and 5 respectively show embodiments of spatial visualizations of a realtime communication session that include visua! cues that reveal the current communication states of two networked communicants involved in the realtime communication session in reiation to props (also referred to as objects) in a graphical representation of a virtual area. Sn these embodiments, the spatial visuaSization includes a graphical representation 46, 48 of each of the communicants in spatial relation to a graphical representation 58 of a virtual area. In particular, the virtual area is represented by a perspective view of a three-dimensional visua! space in which the graphical representations 46, 48 of the communicants can have different respective positions The visualizations shown in FIGS. 4 anά 5 also include props that provide visual cues that reveal the states of various communication channels over which the communicants are configured to communicate, in particular, these visualizations include a viewscreen 60 thai shows the state of application sharing communication sessions, and a table 62 that shows the state of file sharing communication sessions.
[00803 The viewscreen 60 provides visual cues that indicate whether or not a communtcant is sharing an application over an application sharing channel. As shown in FIG. 4. in response to a communicant's selection of the viewscreen 60, the
communicant's sprite 48 automatically is moved to a position in the graphical
representation 58 of the virtual area that is adjacent the viewscreen 60. The position of the communicant's sprite 48 adjacent the viewscreen 60 indicates that the communicant currently is sharing or is about to share an application with the other communicants in the virtual area. The graphical depiction of viewscreen 60 is changed depending on whether or not an active application sharing session is occurring, in the illustrated embodiment, the depicted color of the viewscreen 60 changes from light during an active application sharing session (as shown in FIG. 4) to dark when there is no application sharing taking place (as shown in FiG. 5). Additional details regarding the application sharing process are described in connection with FfGS. 26-28 of U. S Patent Application No. 12/354,709, filed January 15. 2009, and in U.S. Patent Application No. 12/418,270, filed April 3, 2009,
[0061 J The table 62 provides visual cues that indicate whether or not a communicant is sharing or has shared a data file over a data file sharing channel. As shown in FiG. 5, in response to a communicant's selection of the table 62, the communicant's sprite 48 automatically is moved to a position in the graphical
representation 58 of the virtual area that is adjacent the table 82. The position of the communicant's sprite 48 adjacent the viewscreen 60 indicates that the communicant currently is sharing or is about to share a data fiie with the other communicants in the virtual area. In this process, the communicant uploads the data file from the client node 12 to a repository that is maintained by the synchronous conferencing server node 30. In response to the communicant's selection of the data file to upload, the synchronous conferencing server node 30 stores the uploaded fiie in the repository and creates a database record that associates the data file with the table 62. After a data file has been shared by the communicant, the state of the table 62 changes from having a clear table surface (as shown in FlG 4) to having a graphical representation 64 of a data file on the table surface (as shown in FiG. 5). Other communicants in the virtual area 58 are able to view the contents of the uploaded data file by selecting the graphical representation 64 and, subject to governance rules associated with the virtual area 58, optionally may be able to modify or delete the data file. Additional details regarding the file sharing process are described in connection with FiGS, 22 and 23 of U.S. Patent Application No. 12/354,709, filed January 15, 2009.
[00621 FIG. 8 shows an embodiment of a spatia! visualization 70 of two realtime communication sessions in two different virtual areas (i.e., "Virtual Area I" and "Virtual Area Ii"). Each of the virtual areas is represented by a one-dimensional space that contains graphical representations of the communicants who currently have presence in the space. In some embodiments, the ordering of the spatial positions (e.g., from left to right) of the graphical representations of the communicants in each of the virtual areas corresponds to a spatial visualization of the temporal ordering of the communicants in terms of the times when they established respective presences in the virtual areas, in the illustrated embodiments, each communicant is represented by a respective circular sprite 46, 48, 72, 74, 76, 78. The communicant named "Dave" is represented by a respective sprite 48: 78 in each of the virtual areas, reflecting the fact that he is present in both virtual areas. The states of various communication channels over which the respective communicant is configured to communicate are revealed by visual cues that are shown in the spatial visualization 70. For example, the on or off state of a communicant's local speaker channel is depicted by the presence or absence of a headphones graphic 52 on the communicant's sprite. Thus, when the speakers of the communicant who is represented by the sprite are on, the headphones graphic 52 is present (see sprites 46; 48S 72,76, and 78} and, when the communicant's speakers are off, the headphones graphic 52 is absent (see sprite 74). The on or off state of the communicant's microphone is depicted by the presence or absence of a microphone graphic 54 on the communicant's sprite. Thus, when the microphone is on, the microphone graphic 54 is present (see sprites 46 and 72) and; when the microphone is off, the microphone graphic 54 is absent (see sprites 48, 74, 78, and 78). In this way, the headphones graphic 52 and the microphone graphic 54 provide visual cues of the states of the communicant's sound playback and microphone devices. iii. INTEGRATING SPATIAL VISUALIZATIONS WITH LOGS OF REALTIME NETWORKED INTERACTIONS IN A VIRTUAL AREA
A. INTRODUCTION
[00631 Embodiments of the platform are capable of integrating a spatial visualization of realtime networked communications in a virtual area with logs of the interactions that are associated with the virtual area. In this way, current and prior Sogs of communicant interactions are enhanced with references to the spatiai visualization of those interactions, references which engage the communicants' spatial memories of the interactions to enable greater recall and understanding of the contexts of the
interactions.
[0064] in some embodiments, a current realtime communication session is established between communicants operating on respective network nodes. A spatial visualization of the current realtime communication session is displayed on a display. The spattal visualization includes a graphical representation of each of the
communicants in spatial relation to a graphical representation of a virtual area. During the current communication session, a log of event descriptions describing respective events involving interactions of the communicants in the virtual area is presented on the display in contextual association with elements of the spatial visualization of the current realtime communication session.
[0065] In some embodiments , a visual association between respective ones of the event descriptions in the log and elements of the spatial visualization of the current realtime communication session is depicted on the display. For example, a visual association may be depicted between respective ones of the event descriptions in the log and respective ones of the graphical representations of the communicants involved in the events described by the respective event descriptions, In this example, a respective label may be associated with each of one or more of the event
descriptions, where the label has a respective visual appearance that matches a visual element of the graphical representation of the communicant involved in the event described by the event description In this way, the events in the log share a common visual vocabulary with the slate of the communicants in the spatiai visualization shown in the display.
[0068] In some embodiments, in response to an entry of a respective one of the communicants into the virtual area, the graphical representation of the respective communicant is added to the spatial visualization, and a respective one of the event descriptions describing the entry of the respective communicant into the virtual area is presented on the display, in some embodiments, in response to a departure of a respective one of the communicants from the virtυai area, the graphical representation of the respective communicant is removed from the spatsa! visualization, and a respective one of the event descriptions describing the departure of the respective communicant from the virtual area is presented on the display. In some embodiments, in response to a sharing of a data file by a respective one of the communicants with other ones of the communicants, a communicant-seiectabie graphical representation of the data file is displayed in spatial relation to the graphical representation of the virtual area, and a respective one of the event descriptions describing the sharing of the data file by the respective communicant is presented on the display. In some embodiments, in response to a sharing of an application by a respective one of the communicants with other ones of the communicants, a graphical indication of the sharing of the application in spatial relation to the graphicai representation of the virtual area is displayed on the display, and a respective one of the event descriptions describing the sharing of the application by the respective communicant is displayed on the display.
[0067J FlG. 7 shows an embodiment of a method by which the platform integrates spatiai visualizations of realtime networked interactions in a virtual area with historical records of the interactions that are associated with the virtual area.
[0088] in response to the initiation of a current realtime communication session in a virtual area (FIG. 7, block 80), the platform retrieves context configuration data that includes a log of interactions that are associated with the virtual area (FlG, 7, block 82). The tog typically includes data that is extracted from the interaction records 38, which describe the contexts of interactions between communicants in the virtual area. The extracted data may include, for example, data stream data (e.g., text chat entries) and references (e.g., hyperlinks) to files anά data streams (e,g.: audio and video data streams) that are shared or recorded during one or more prior
communication sessions in the virtual area.
[0089] The platform generates a visualization of the current realtime communication session in the virtual area in association with the historical log (FIG. 7, block 84). In this process, the platform typically retrieves context data describing an end state of the preceding communications session in the virtual area: including the positions and states of the props in the virtual area. The spatial visualization that is generated includes a graphical representation of each of the communicants in spatial relation to a graphical representation of the virtual area. The virtual area may be represented graphically by any type of one-dimensional, two-dimensional, or three- dimensional view that situates the graphical representations of the communicants in respective positions in a visual space. During the current communication session, the platform depicts visual cues in the spatial visualization that shows current
communication states of the communicants. Each of the communication states typically corresponds to a state of a respective communication channel (e.g., text chat, audio, video, application share, and file share channel) over which a respective one of the communicants is configured to communicate,
[0070] During the current realtime communication session, the platform stores context configuration data that includes records of interactions between the
communicants that occur in the virtual area; where the records are indexed by an identifier of the virtual area (FIG. 1, block 86) Each interaction record describes the context of an interaction between a pair of the communicants in the virtual area. For example, in some embodiments, an interaction record contains an identifier for each of the communicants, an identifier for the place of interaction (e.g., a virtual area instance), a description of the hierarchy of the interaction place (e.g., a description of how the interaction area relates to a larger area), start and end times of the interaction, and a list of all files and other data streams that are shared or recorded during the interaction. Thus, for each realtime interaction, the interaction platform tracks when it occurred, where it occurred, and what happened during the interaction in terms of communicants involved (e g., entering and exiting), objects that are activated/deactivated, and the files that were shared.
[00713 in response to the termination of the current communication session (FlG. 7, block 88), the platform stores context configuration data that describes the end state of the current communication session (FIG. 7, block 90). The enά state context configuration data typically includes a description of all props (e.g., viewscreen and table props) that are present in the virtual area at the time the current communication session was terminated, including a description of the positions of the props and their respective states (e.g., associations between a table prop and data files that were shared in the virtual area). The enά state context configuration data typically is used by the platform to recreate the end slate of the virtual area for the next realtime communication session that takes place in the virtual area,
B. EXEMPLARY SPATIAL INTERFACES FOR REALTIME CHAT
INTERACTIONS
[QQ72J Some embodiments apply one or more of the spatial metaphor visualizations described above on top of realtime chat interactions. These visualizations provide a context for depicting the current communication states of the communicants involved in realtime chat interactions. The spatial metaphor also provides a context for organizing the presentation of various interface elements that are used by
communicants to participate in realtime chat interactions. The spatial metaphor visualizations may be applied to any type of instant messaging platform that provides realtime text-based communication between two or more communicants over the internet or some form of interna! network/intranet, optionally with one or more other realtime communication channels, such as audio, video, file share, and application sharing channels. For example, embodiments may be integrated with any of the currently available instant messaging platforms including, for example, AOL Instant Messenger, MSN Messenger, Yahoo! Messenger, Google Talk, and Skype.
[0073] FIG. 8 shows an exemplary embodiment of a spatial interface 92 for a realtime chat interaction between a group of communicants in a virtual area. Each of the communicants is represented graphically by a respective sprite 94, 98, 98, 100, 102 and the virtual area is represented graphically by a two-dimensional top view of a rectangular space 101 (i.e., the "West Conference" space). When the communicants initially enter the virtual area, thetr sprites automatically are positioned in predetermined locations (or "seats") in the virtual area. The virtual area includes two viewscreen props 104, 106 and a table prop 108, Communicants interact with the props by selecting them with a input device (e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like).
[0074J The spatial interface 92 is integrated with a realtime communications interface window 110 that also includes a toolbar 112, a chat log area 114, a text box 1 16, and a Send button 118, The user may enter text messages in the text box 116 and transmit the text messages to the other communicants in the currently West Conference space 101 by selecting the Send button 118. The spatial interface 92 and the chat iog area 114 are separated by a splitter 117 that, in some embodiments, can be slid up and down by the user to hide or reveal the spatial interface 92.
[0075] The °haf log area 114 displays a log of current and optionally prior events that are associated with the West Conference space 101 , An exemplary set of events that are dispiayed in the chat iog area 114 include: text messages that the user has exchanged with other communicants in the West Conference space 101 ; changes in the presence status of communicants in the West Conference space 101 ; changes in the speaker and microphone settings of the communicants in the West Conference space 101 ; and the status of the props 104-108, inciuding references to any
applications and data files that are shared in connection with the props. In the illustrated embodiments, the events are labeled by the communicant's name followed by content associated with the event (e.g., a text message) or a description of the event. For example, in the example shown in FlG. 8, status related events are labeled as follows;
• $UserName$ entered the room.
• $UserName$ left the room.
• $UserName$ shared $ProcessName$ on $ViewScreenName$,
• $UserName$ cleared $ViewScreenName$ where the tags between "$" and "$" identify communicants, shared applications, or props. In addition, each of the events is associated with a respective timestamp 119 that identifies the date and time when the associated event was initiated.
[00763 in embodiments that are integrated with conventional instant messaging platforms (e.g., AOL instant Messenger, MSN Messenger, Yahoo?
Messenger, Google Talk, and Skype), the chat log area 114 typically contains a standard "chat history" (also referred to as an "instant message history") that includes a list of entries typed remotely by two or more networked communicants, interleaved in the order the entries have been typed. The chat history typically is displayed on each communicant's terminal display, along with an indication of which user made a particular entry and at what time relative to other communicant's entries. This provides a session history for the chat by enabling communicants to independently view the entries and the times at which each entry was made. [0077] The spatial visualization 92 provides a context for organizing the presentation of the events that are displayed in the chat log area 114, For example, in the illustrated embodiment, each of the displayed events is labeled with a respective tag that visually correlates with the appearance of the sprite of the communicant that sourced the displayed event. In particular, each of the events that is sourced by a particular one of the communicants is labeled with a respective icon 130, 132, 134, 136 with a visual appearance {e.g., color-code) that matches the visual appearance of that communicant's sprite, in this example, the color of the icons 130, 134 matches the color of the body of Dave's sprite 100. the color of the icon 132 matches the color of the body of Camilla's sprite 98, and the color of the icon 136 matches the color of the body of Jack's sprite 96,
[0078] The toolbar 112 includes a set of navigation and interaction control buttons, including a headphones button 120 for toggling on and off the user's speakers, a microphone button 122 for toggling on and off the user's microphone, a get button 124 for getting people, a map button 126 for opening a map view of a larger virtual area the contains the space 101 , and a reconnect button 128 for reestablishing a connection to the virtual area.
[0079] After the user has moved into the West Conference space 101 , the user may toggle one or both of the headphones button 120 and the microphone button 122 in order to selectively turn-on and turn-off one or both of the user's speakers and microphone. As explained above, the headphones graphic, the radiating concentric circles around the user's sprite, and the microphone graphic on the user's sprite are omitted when the user's speakers and microphone both are turned-off.
[0080] Referring to FSG. 9r in response to a user selection of the get button 124, a list of communicants is displayed in a separate frame 138. The communicants are segmented into two groups: a first group labeled "People in West Conference" that identifies all the communicants who are in the current area (i.e., West Conference); and a second group labeled "Lansing Aviation" that identifies all the communicants who are present in a larger area (i.e., Lansing Aviation, which contains the current area) but are not present in the current area. Each of the virtual areas is represented by a respective one-dimensional space 142, 144 that contains graphical representations of the communicants who currently have presence in the space. In some embodiments, the ordering of the spatial positions (e.g.: from top to bottom) of the graphical representations of the communicants in each of the virtual areas 142 144 corresponds to a spatial visualization of the temporal ordering of the communicants in terms of the times when they established respective presences in the virtual areas, Sn the illustrated embodiments, each communicant is represented by a respective circular sprite that is labeled with a respective user name of the communicant (i.e., "Jack," "Dave," "Camilla:' "Karou/' "Arkadi/1 Υuka," "Teca/' Υoshi/' and "Adam1").
[00813 The states of various communication channels over which the respective communicant is configured to communicate are revealed by visual cues that are shown in the spatiai visualizations of the communicants in the virtual areas 142, 144, For example, the on or off state of a communicant's loca! speaker channel is depicted by the presence or absence of a headphones graphic 52 on the
communicant's sprite. Thus, when the speakers of the communicant who is
represented by the sprite are on, the headphones graphic 52 is present (see sprites Jack, Dave, Camilla, Karou, Arkadi, and Teca) and, when the communicant's speakers are off, the headphones graphic 52 is absent (see sprites Yuka: Yoshl and Adam). The on or off state of the communicant's microphone is depicted by the presence or absence of a microphone graphic 54 on the communicant's sprite. Thus, when the microphone is on, the microphone graphic 54 is present (see sprites Karou and Teca) and, when the microphone is off, the microphone graphic 54 is absent (see sprites Jack, Dave, CamiSla, Arkadi, Yuka, Yoshi, and Adam}. (The radiating circles that indicate the on state of a communicant's microphone graphic typically is omitted in this
visualization,} The headphones graphic 52 and the microphone graphic 54 provide visual cues of the states of the communicant's sound playback and microphone devices. The activity state of a communicant's text chat channel is depicted by the presence or absence of the hand graphic 57 adjacent the communicant's sprite (see sprite Adam). Thus, when a communicant is transmitting text chat data to another network node the hand graphic 57 is present, and when a communicant is not transmitting text chat data the hand graphic 57 is not present. In some embodiments, text chat data is transmitted only when keyboard keys are depressed, in which case the visualization of the communicant's text channel appears as a flashing on and off of the hand graphic 57
[0082! *n response to a user selection of one of the communicants in the list of available communicants in the frame 138, the platform transmits an invitation to the selected communicant to join the user in the respective zone. For example, FIG. 10 shows a pop-up window 141 that is generated by the piatform in the situation in which the user has selected "Arkadi" in the iist of available communicants displayed in the frame 138. in response to the selection of the Send button 143, the platform transmits an invitation to the communicant who is associated with the name Arkadi to join the user in the West Conference space 101 (e.g. , "Please join me in West Conference ~ Jack."}.
C. EXEMPLARY SPATIAL INTERFACES FOR PRIVATE REALTIME
NETWORKED INTERACTIONS
[QQ83J Some embodiments apply one or more of the spatial metaphor visualizations described above on top of realtime private interactions between (typically only two) networked communicants. These spatial visualizations enable the depiction of a current private realtime communications session between the communicants in the context of their prior private relationship history. In other words, the semantics of the virtual area is the relationship history between the communicants. The spatial visualizations also provide a framework for organizing the presentation of various interface elements that are used by communicants to participate in private realtime networked communications in the context of their prior relationship history.
[0084] A current private realtime communications session between
communicants typically is visualized as a private virtual area that provides a reference for the records of the private interactions that occur in the private virtual area, records which are stored persistently in the relationship database 36 in association with the private virtual area. The virtual area typically is created automatically during the first communication session and then persists until one or all of the communicants choose to delete it. By default, the private virtual area typically is owned jointly by all the
participating communicants. This means that any of the communicants can freely access the private virtual area and the associated private interaction records, and can unilaterally add. copy, or delete the private virtual area and all the associated private interaction records.
[QQ85J Each communicant typically must explicitly navsgate to the private virtual area that he or she shares with another communicant, fn some embodiments, this is achieved by selecting an interface control that initiates a private communication with the other communicant. For example, in some embodiments, in response to the initiating of a private instant messaging communication (e.g., a text, audio, or video chat) with another communicant, the platform automatically situates the private communication in a private virtual area that typically is configured in accordance with configuration data that describes the prior state of the private virtual area when the communtcants last communicated in the private virtual area.
[00881 *n some embodiments, the platform responds to the receipt of a command from a first communicant operating on a first network node to initiate a private communication with a second communicant operating on a second network node as foliows. The platform estabiishes a current realtime communication session between the first and second network nodes. The piatform identifies a private virtual area that is associated with the first and second communicants, The piatform retrieves context configuration data associated with the private virtual area and generated in response to interactions of the first and second communicants in the private virtual area. On a display, the piatform displays a spatiaϊ visualization of the current realtime
communication session, where the spatial visualization includes graphical
representations of the first and second communicants in spatial relation to a graphical representation of the virtual area configured in accordance with the context
configuration data.
[0087J In some embodiments, during the current realtime communication session, the platform generates a log of event descriptions describing respective events involving interactions of the first and second communicants in the virtual area. During the current realtime communication session, piatform typically stores the event descriptions in a data storage device with an index comprising an identifier of the virtual area. The log of event descriptions may include, for example, at least one of: text of a chat conversation between the first and second communicants in the virtual area; a description of a data file shared by a respective one of the first and second
communicants in the virtuai area; and a description of an application shared by a respective one of the first and second communicants in the virtual area. During the current realtime communication session, typically presents the log of event descriptions on the display. The log of event descriptions typically is presented in contextual association with elements of the spatial visualization of the current realtime
communication session,
[00881 *n some embodiments, the platform retrieves context configuration data that includes a iog of event descriptions describing respective events involving interactions of the first and second communicants in the virtual area during one or more prior communication sessions before the current communication session. The platform typicaily presents the tog of event descriptions generated during the current realtime communication session together with the retrieved context configuration data
comprising the log of even! descriptions.
[0089] In some embodiments, the platform retrieves context configuration data that includes a description of an end state of a prior realtime communication session between the communicants and dispiays the graphical representation of a virtual area in a state that corresponds to the end state of the prior communication session between the communicants.
[0090] FiG. 11 shows an embodiment of a method of managing realtime networked communications between networked communicants in a private virtual area. In response to a determination that a private realtime communication between communicants has been initiated (FlG. 11 , block 150), the platform determines whether or not a private virtual area that is indexed by the identifiers of all the communicants already has been created (FiG. 11 , block 152). If such a private virtual area already has been created, the platform retrieves a specification of the private virtual area (FlG, 11 ; block 154); the platform also, retrieves context configuration data that is associated with the private virtual area (FSG. 11 , block 156). If a private virtual area that is indexed by the identifiers of all the communicants has not already been created, the platform creates a new private virtual area that is indexed by identifiers of ail the communicants (FSG. 11 , block 158). After the specification of the private virtual area has been either retrieved or newly created, the platform generates a visualization of the current realtime communication session in the private virtual area configured in its current context (i.e., either in its prior configuration or in its new default configuration) (FIG. 11 , block 160), During the current private realtime communication session, the platform stores context configuration data that describes the state of the private virtual area and includes records of interactions in the private virtual area, which records are indexed by the identifier of the private virtual area (FIG. 11 , block 162),
[0091 ] FIG. 12 shows an embodiment of a process 168 of generating a spatial visualization of a current realtime communication session. In this process, each of the communicants (A and B) is represented by a respective node 170, 172 anά their private bilateral relationship ts represented by an edge 174 of a graph that interconnects the nodes 170, 172. The bilateral relationship between the communicants is defined by their interaction history in the private virtual area. The interaction history is stored in the interaction database 38 in the form of interaction records that describe the interactions of the communicants in the private virtual area. These interactions can include any of the interactions involving any of the communication channels over which the
communicants are configured to communicate, including, for example, chat, audio, video, realtime differential streams of tagged records containing configuration
instructions, 3D rendering parameters, and database query results (e.g., streams keyboard event streams relating to widget state changes, mouse event streams relating to avatar motion, and connection event streams), application sharing, file sharing, and customizations to the private virtual area, in the illustrated embodiment, the interaction history between the communicants is integrated with a template 178 that describes a graphical representation of the private virtual area to produce the spatial visualization 180 of the current realtime communication session. In this process, the private virtual area is configured in accordance with the customization records in the interaction history. The private virtual area also is populated with the other elements of the interaction history in accordance with the specification provided by the template 178.
[0092J FlG. 13 shows an embodiment of a data model 180 that relates private virtual area identifiers to communicants, template specifications, and context data. Sn accordance with this data model 180, each private virtual area is associated with a respective unique identifier (e.g., AreaJDI and ArβaJD2) and is indexed by the respective identifiers (e.g., ComrnJDA, CommJDB, CommJDX, and ComrnJDY) of all the communicants who own the private virtual area, In the examples shown in FIG, 13, each of the private virtual areas is jointly owned by a respective pair of communicants Each area identifier is associated with a respective template specification identifier that uniquely identifies a particular area specification. Each area identifier also is associated with a respective configuration data identifier that uniquely identifies a particular set of data (e.g., customization data) that is used by the platform to configure the private virtual area.
[0093] PiG. 14 shows an embodiment of a data model 182 that relates interaction records 38 in the relationship database 36 with respective ones of the private virtual areas. This relationship is used by the platform in the process of populating the private virtual area with the elements of the interaction history in accordance with the associated template specification.
[0094] FiGS. 15 and 16 show an embodiment of a spatia! interface 188 for realtime networked communications between communicants in a private virtual communication area (labeled "Chat with Dave") that ss created by the platform for the private bilateral interactions between the user (i.e.. Jack) and another communicant (i.e., Dave). FIG. 15 depicts an exemplary state of the private virtual area in which Dave left the area after having just interacting with Jack, who still is in the private virtual area. FIG. 16 depicts the state of the private virtual area in which Jack just entered the area, which already was occupied by Dave.
[0095] The spatial interface 188 provides a spatial visualization of the private virtual area. In this visualization, each of the communicants is represented graphically by a respective sprite 196, 198 and the private virtual area is represented graphically by a 2.5-dimensional iconographic view of a cloud. The iconographic cloud view
distinguishes the private virtual area from other types of virtual areas in a way that reinforces the notion that the focus of the private virtual area, first and foremost, is the relationship between the communicants as opposed to the area. Sn contrast, other types of virtual areas (e.g., West Conference), the centra! focus typically relates to matters that traditionally are associated with real-world physical spaces (e.g., work, home, meetings, clubs, etc.),
[0096] When the communicants initially enter the private virtual area, their sprites automatically are positioned in predetermined locations (or "seats") in the private virtual area. In the illustrated embodiment, the private virtual area includes a
viewscreen prop 200. In this embodiment, in response to the selection of the
viewscreen object 200, the graphical representation of a communicant is repositioned adjacent to the viewscreen object and a pair of eyes is added to the graphical representation to provide an additional visual indication that the associated
communicant is viewing an application in connection with the viewscreen object 200.
[00973 The communicants that are associated with the private virtual area may customize the private virtual area, for example, by adding additional props (e.g., another viewscreen prop or a table prop), changing the color scheme, etc.
Communicants interact with the props by selecting them with a input device (e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like), in response to a communicant's selection of a particular prop, the communicant's sprite either is repositioned adjacent to the selected prop or it is replicated and the replicated sprite is positioned adjacent to the selected prop and the original sprite remains where it was seated.
[00981 f^e spatial interface 188 is integrated with a realtime communications interface window 190 that additionally includes a toolbar 192, a chat log area 194 a text box 206, and a Send button 208 that function in the same way as the toolbar 112, the chat log area 114, the text box 116, and the Send button 118 of the spatial interface 110 shown in FIG, 8,
[00991 The chat log area 194 displays a log of events that are associated with the private bilateral interactions between the user (i.e., Jack) and another one of the communicants (i.e., Dave). The log of events includes sequences of text messages that the user has exchanged with the other communicant in the associated private virtual area. The user may enter text messages in the text box 206 and transmit the text messages to the other communicant in the private virtual area by selecting the Send button 208. An exemplary set of events that can be recorded in the chat log area 204 include; text message entries; changes in the presence status of communicants in the private virtual area; changes in the speaker and microphone settings of the
communicants in the private virtual area; and the status of any props (e.g., viewscreen 200). including references to any applications and data files that are shared in
connection with the props.
[00100] in the illustrated embodiments, the events are labeled by the
communicants' names followed by content associated with the event (e.g., a text message) or a description of the event. In FiGS. 15 and 18, status related events are labeled as follows:
• $UserName$ entered the room,
• $UserName$ left the room.
• $UserName$ shared $ProcessName$ on $ViewScrβenName$.
• $UserName$ cleared $ViewScreenName$ where the tags between '$" and "$" identify communicants, shared applications, or props, in addition, each of the events is associated with a respective timestamp 209 that identifies the date and time of the associated event. In another example, the application sharing event description 214 has a description of the event class (Share), the identity of the sharer (Dave), the label of the share target (Screen 1 ), the URL of the share target (represented by the underlining of the share target label), the timestamp associated with the event, and a description of the shared application.
[001013 As shown in FIG. 18, a graphics! separator, such as ruie line 216, is added to the chat log area 194 between the events of one communication session (also referred to as a "conversation"') and those of another communication session. In some embodiments, the textuai descriptions of prior communication sessions are
deemphasized (e.g., by using a lighter font color, such as gray) so that the events that area associated with the current communication session stand out visually.
[001021 in some embodiments, previous conversations are "collapsed" and labeled with the iist of participants in the conversation as weli as a timestamp of the most recent event or message within the conversation. Clicking a "toggie" to the Seft of the conversation iabe! opens up the conversation and displays the fuSl contents of the conversation in the chat log area 194
[00103] in embodiments that are integrated with conventionai instant messaging platforms (e.g., AOL Instant Messenger, MSN Messenger, Yahoo!
Messenger, Googie Taik, and Skype), the chat tog area 194 contains a standard "chat history" (also referred to as an "instant message history") that includes a list of entries typed remotely by two or more networked communicants, interleaved in the order the entries have been typed. The chat history typically is displayed on each communicant's terminal display, along with an indication of which user made a particular entry and at what time reiative to other communicant's entries. This provides a session history for the chat by enabling communicants to independently view the entries and the times at which each entry was made.
[00104] The spatial interface 188 provides a context for organizing the presentation of the events that are displayed in the chat log area 194. For example, in the illustrated embodiment, each of the displayed events is labeled with a respective tag that visualiy correlates with the appearance of the sprite of the communicant that sourced the displayed event, in particular, each of the events that is sαurced by a particular one of the communicants is labeled with a respective icon 210, 212 with a visual appearance (e.g., color-code) that matches the vssuai appearance of that communicant's sprite, in the illustrated embodiment, for example, the color of the icon 212 matches the color of the body of Dave's sprite 198 and the color of the icon 210 matches the color of Jack's sprite 196.
[00105] FIG. 17 shows an embodiment of a spatial interface 220 for realtime networked communications between communicants in a private virtual area (labeled "Chat with Yuka") that is created by the platform for the private bilateral interactions between the user (i.e., Arkadi) and another communicant (i.e., Yuka). The spatial interface 220 provides a spatial visualization of the private virtual area, in this
visualization , each of the communicants is represented graphically by a respective sprite 222, 224 and the virtual area is represented graphically by a 2.5-dimensional iconographic view of a cloud. The spatial interface 220 is integrated with a realtime communications interface window 218 that additionally has the same interface elements as the interface window 190 shown in FIGS. 15 and 16, including a toolbar 192, a chat log area 194, a text box 206, and a Send button 208.
[00106] When the communicants initially enter their private virtual area, their sprites automatically are positioned in predetermined locations (or "seats") in the private virtual area. In the illustrated embodiment, the private virtual area includes two viewscreen props 226, 228 and a table prop 230, on top of which is shown a graphical representation 231 of a data fife (i.e., "DE Expense Report__ml.doc") that was shared by a respective one of the communicants. The communicants that are associated with the private virtual area may customize the pnvate virtual area, for example, by adding additional props (e.g., another viewscreen prop or a table prop), changing the color scheme, etc. Communicants interact with the props by selecting them with an input device (e.g., by double-clicking on the props with a computer mouse, touch pad, touch screen, or the like). In response to a communicant's selection of a particular prop, the communicant's sprite either is repositioned adjacent to the selected prop or it is replicated and the replicated sprite is positioned adjacent to the selected prop and the original sprite remains where it was seated, In the example shown in FIG. 17, Yuka has selected the viewscreen 228 and, in response, the platform has created a copy 232 of her original sprite 224 at a location adjacent the selected viewscreen 228, While an application (or process) is being shared, the viewscreen 228 is shown to be in an active state, which is visually distinguishable from the depiction of the inactive viewscreen 226. IV. EXEMPLARY SYSTEM ARCHITECTURE
A INTRODUCTION
[00107] FiG. 18 is a diagrammatic view of an embodiment 300 of the network communication environment 10 (see FiG. 1 ) in which the synchronous conferencing server node 30 is implemented by a virtual environment creator 302, The virtual environment creator 302 includes at least one server network node 304 that provides a network infrastructure service environment 306. The communications application 26 and the network infrastructure service environment 306 together provide a platform for creating a spatial virtual communication environment (also referred to herein simply as a "virtual environment" } that includes one or more of the spatia! metaphor visualizations described above.
[00108] The network infrastructure service environment 306 manages sessions of the first and second client nodes 12, 14 in a virtual area 308 in accordance with a virtual area application 310. The virtual area application 310 is hosted by the virtual area 308 and includes a description of the virtual area 308. The communications applications 26 operating on the first and second client network nodes 12, 14 present respective vtews of the virtual area 308 in accordance with data received from the network infrastructure service environment 306 anά provide respective interfaces for receiving commands from the communicants and providing a spatial interface that enhances the realtime communications between the communicants as described above. The communicants typically am represented in the virtual area 308 by respective avatars, which typically move about the virtual area 308 in response to commands that are input by the communicants at their respective network nodes. Each communicant's view of the virtual area 308 typically is presented from the perspective of the communicant's avatar, which increases the level of immersion experienced by the communicant. Each communicant typically is able to view any part of the virtual area 308 around his or her avatar. In some embodiments, the communications applications 26 establish realtime data stream connections between the first and second client network nodes 12, 14 and other network nodes sharing the virtual area 308 based on the positions of the communicants' avatars in the virtual area 308.
[00109] The network infrastructure service environment 306 also maintains the relationship database 36 that contains the records 38 of interactions between communicants. Each interaction record 38 describes the context of an interaction between a pair of communicants.
B. NETWORK ENViRONMENT
[00110] The network 18 may include any of a local area network (LAN), a metropolitan area network (MAN)1 and a wide area network (WAN) (e.g., the internet). The network 18 typically includes a number of different computing platforms and transport facilities that support the transmission of a wide variety of different media types (e.g.. text, voice, audio, and video) between network nodes,
[001111 The communications application 26 (see FIGS. 1 and 18) typically operates on a client network node that includes software and hardware resources which, together with administrative policies, user preferences (including preferences regarding the exportation of the user's presence and the connection of the user to areas and other users), and other settings, define a local configuration that influences the administration of realtime connections with other network nodes. The network connections between network nodes may be arranged in a variety of different stream handling topologies, including a peer-to-peer architecture, a server-mediated
architecture, and hybrid architectures that combine aspects of peer-to-peer and server- mediated architectures. Exemplary topologies of these types are described in US Application Nos. 1 1/923,629 and 1 1/923,634, both of which were filed on October 24, 2007,
C. NETWORK INFRASTRUCTURE SERVICES
[00112] The network infrastructure service environment 30 typically includes one or more network infrastructure services that cooperate with the communications applications 26 in the process of establishing and administering network connections between the client nodes 12, 14 and other network nodes (see FIGS, 1 and 18), The network infrastructure services may run on a single network node or may be distributed across multiple network nodes. The network infrastructure services typically run on one or more dedicated network nodes (e.g., a server computer or a network device that performs one or more edge services, such as routing and switching) In some embodiments, however, one or more of the network infrastructure services run on at least one of the communicants' network nodes. Among the network infrastructure services that are included in the exemplary embodiment of the network infrastructure service environment 30 are an account service, a security service, an area service, a rendezvous service, and an interaction service.
Account Service
[00113] The account service manages communicant accounts for the virtual environment. The account service also manages the creation and issuance of authentication tokens that can be used by client network nodes to authenticate themselves to any of the network infrastructure services.
Security Service
[00114] The security service controls communicants1 access to the assets and other resources of the virtual environment. The access contro! method implemented by the security service typically is based on one or more of capabilities (where access is granted to entities having proper capabilities or permissions) and an access control list (where access is granted to entities having identities that are on the list). After a particular communicant has been granted access to a resource, that communicant typically uses the functionality provided by the other network infrastructure services to interact in the network communications environment 300.
Area Service
[00115} The area service administers virtual areas. In some embodiments, the area service remotely configures the communications applications 26 operating on the first and second client network nodes 12, 14 in accordance with the virtual area application 308 subject to a set of constraints 312 (see FlG. 18). The constraints 312 typically include controls on access to the virtual area. The access controls typically are based on one or more of capabilities (where access is granted to communicants or client nodes having proper capabilities or permissions) and an access control list (where access is granted to communicants or client nodes having identities that are on the list).
[001 16] The area service also manages network connections that are associated with the virtual area subject to the capabilities of the requesting entities, maintains global state information for the virtual area, and serves as a data server for the client network nodes participating in a shared communication session in a context defined by the virtual area 308. The global state information includes a list of ail the objects that are in the virtual area and their respective locations in the virtual area. The area service sends instructions that configure the client network nodes. The area service also registers and transmits initialization information to other client network nodes that request to join the communication session. In this process, the area service may transmit to each joining client network node a list of components (e.g., plugins) that are needed to render the virtual area 308 on the client network node in accordance with the virtual area application 310. The area service also ensures that the client network nodes can synchronize to a global state if a communications fault occurs. The area service typically manages communicant interactions with virtual areas via governance rules that are associated with the virtual areas.
Rendezvous Service
[001 17] The rendezvous service manages the collection, storage, and distribution of presence information and provides mechanisms for network nodes to communicate with one another (e.g., by managing the distribution of connection handles) subject to the capabilities of the requesting entities. The rendezvous service typically stores the presence information in a presence database. The rendezvous service typically manages communicant interactions with each other via communicant privacy preferences.
Interaction Seryjce
[00118] The interaction service maintains the relationship database 36 that contains the records 38 of interactions between communicants. For every interaction between communicants, one or more services of the network infrastructure service environment 306 (e.g., the area service) transmit interaction data to the interaction service. In response, the interaction service generates one or more respective interaction records and stores them in the relationship database. Each interaction record describes the context of an interaction between a pair of communicants. For example, in some embodiments, an interaction record contains an identifier for each of the communicants, an identifier for the place of interaction (e.g., a virtual area instance), a description of the hierarchy of the interaction place (e.g., a description of how the interaction room relates to a larger area), start and end times of the interaction, and a list of all files and other data streams that are shared or recorded during the interaction. Thus, for each realtime interaction, the interaction service tracks when it occurred, where it occurred, and what happened during the interaction in terms of communicants involved {e.g., entering and exiling), objects that are activated/deactivated, and the files that were shared.
[00119] The interaction service also supports queries on the relationship database 36 subject to the capabilities of the requesting entities. The interaction service presents the results of queries on the interaction database records in a sorted order (e.g.. most frequent or most recent) based on virtual area. The query results can be used to drive a frequency sort of contacts whom a communicant has met in which virtual areas , as well as sorts of who the communicant has met with regardless of virtual area and sorts of the virtual areas the communicant frequents most often. The query results also may be used by application developers as part of a heuristic system that automates certain tasks based on relationships. An example of a heuristic of this type is a heuristic that permits communicants who have visited a particular virtual area more than five times to enter without knocking by default, or a heuristic that allows
communicants who were present in an area at a particular time to modify and delete files created by another communicant who was present in the same area at the same time. Queries on the relationship database 36 can be combined with other searches. For example, queries on the relationship database may be combined with queries on contact history data generated for interactions with contacts using a communication system (e.g., Skype, Facebook, and Flickr) that is outside the domain of the network infrastructure service environment 306.
D. VIRTUAL AREAS
[001201 ~^e communications application 26 and the network infrastructure service environment 306 typically administer the realtime connections with network nodes in a communication context that is defined by an instance of a virtual area. The virtual area instance may correspond to an abstract (non-geometric) virtual space that is defined with respect to abstract coordinates. Alternatively, the virtual area instance may correspond to a visual virtual space that is defined with respect to one-, two- or three- dimensional geometric coordinates that are associated with a particular visualization. Abstract virtual areas may or may not be associated with respective visualizations, whereas visual virtual areas are associated with respective visualizations.
[00121 ] As explained above, communicants typically are represented by respective avatars (e.g., sprites) in a virtual area that has an associated visualization The avatars move about the virtual area in response to commands that are input by the communicants at their respective network nodes. In some embodiments, the communicant's view of a virtual area instance typically is presented from the
perspective of the communicant's avatar, and each communicant typically is able to view any part of the visual virtual area around his or her avatar, increasing the level of immersion that is experienced by the communicant
[00122] A virtual area typicaily includes one or more zones that are associated with respective rules that govern the switching of realtime data streams between the network nodes that are represented by the avatars in the virtual area. The switching rules dictate how local connection processes executing on each of the network nodes establishes communications with the other network nodes based on the locations of the communicants3 avatars in the zones of the virtual area. A virtua! area typically is defined by a specification that includes a description of geometric elements of the virtual area and one or more rules, including switching rules and governance rules. The switching rules govern realtime stream connections between the network nodes. The governance ruies controi a communicant's access to resources, such as the virtual area itself, regions with the virtual area; and objects within the virtual area. In some embodiments, the geometric elements of the virtua! area are described in accordance with the COLLADA - Digital Asset Schema Release 1.4.1 April 2006 specification {availabie from http://www.khronos.org/coilada/), and the switching ruies are described using an extensible markup language (XML) text format (referred to herein as a virtual space description format (VSDL)) in accordance with the COLLADA Streams Reference specification described in US Application Nos. 11/923,629 and 11/923;634.
[001231 ~^e geometric eiements of the virtua! area typically include physical geometry and collision geometry of the virtual area. The physical geometry describes the shape of the virtual area. The physical geometry typically is formed from surfaces of triangles, quadrilaterals, or polygons. Colors and textures are mapped onto the physical geometry to create a more realistic appearance for the virtua! area. Lighting effects may be provided, for example, by painting Sights onto the visual geometry and modifying the texture, color, or intensity near the lights. The coilision geometry describes invisible surfaces that determine the ways in which objects can move in the virtual area. The collision geometry may coincide with the visual geometry, correspond to a simpler approximation of the visuai geometry, or relate to application-specific requirements of a virtual area designer. [00124] The switching rules typically include a description of conditions for connecting sources and sinks of realtime data streams in terms of positions in the virtual area. Each rule typically includes attributes that define the realtime data stream type to which the rule applies and the iocation or locations in the virtual area where the rule applies. In some embodiments, each of the rules optionally may include one or more attributes that specify a required role of the source, a required role of the sink, a priority level of the stream, and a requested stream handling topology. In some embodiments, if there are no expiicit switching rules defined for a particular part of the virtual area, one or more implicit or default switching rules may apply to that part of the virtual area. One exemplary default switching rule is a rule that connects every source to every
compatible sink within an area: subject to policy rules. Policy ruies may apply globally to all connections between the client nodes or only to respective connections with individual client nodes. An example of a policy rule is a proximity policy rule that only allows connections of sources with compatible sinks that are associated with respective objects that are within a prescribed distance (or radius) of each other in the virtual area.
[00125] In some embodiments, governance rules are associated with a virtual area to control who has access to the virtual area, who has access to its contents, what is the scope of that access to the contents of the virtual area (e.g., what can a user do with the contents), and what are the follow-on consequences of accessing those contents (e.g., record keeping, such as audit logs, and payment requirements), In some embodiments, an entire virtual area or a zone of the virtual area is associated with a "governance mesh." Sn some embodiments, a governance mesh is implemented in a way that is analogous to the implementation of the zone mesh described in US
Application Nos. 1 1/923,62S and 1 1/923,634. A governance mesh enables a software application developer to associate governance rules with a virtual area or a zone of a virtual area. This avoids the meά for the creation of individual permissions for every file in a virtual area and avoids the need to deal with the complexity that potentially could arise when there is a need to treat the same document differently depending on the context.
[00126] In some embodiments, a virtual area is associated with a governance mesh that associates one or more zones of the virtual area with a digital rights management (DRM) function. The DRIVI function controls access to one or more of the virtual area or one or more zones within the virtual area or objects within the virtual area. The DRM function is triggered every time a communicant crosses a governance mesh boundary within the virtual area. The DRM function determines whether the triggering action is permitted and, if so, what is the scope of the permitted action, whether payment is needed, and whether audit records need to be generated. !n an exemplary implementation of a virtual area, the associated governance mesh is configured such that if a communicant is able to enter the virtual area he or she is able to perform actions on all the documents that are associated with the virtual area, including manipulating the documents, viewing the documents, downloading the documents, deleting the documents, modifying the documents and re-upioading the documents. In this way, the virtual area can become a repository for information that was shared and discussed in the context defined by the virtual area
[00127] Additional details regarding the specification of a virtual area are described in US Application Nos. 61/042714 (which was filed on April 4, 2008)
11/923,629 (which was filed on October 24, 2007), and 11/923,634 (which was filed on October 24, 2007)
E, COMMUNICATIONS APPLICATION
[00128] in some embodiments, the communications application 26 includes: a, local Human Interface Devices (HiDs) and audio playback devices: b, a So3D graphical display, avatar, and physics engine; c, a system database and storage facility.
1. LOCAL HUMAN INTERFACE DEVICES (HiDS) AND AUDIO PLAYBACK DEVICES
[00129] The local HIDs enable a communicant to input commands and other signals into the client network node while participating in a virtual area communications session. Exemplary HiDs include a computer keyboard, a computer mouse, a touch screen display, and a microphone
[00130] The audio playback devices enable a communicant to playback audio signals that are received during a virtual area communications session. Exemplary audio playback devices include audio processing hardware (e.g., a sound card) for manipulating (e.g., mixing and applying special effects) audio signals, and speakers for outputting sounds. 2. SO3D GRAPHICAL DISPLAY, AVATAR AND PHYSICS ENGiIME
[001311 The So3D engine is a three-dimensional visualization engine that controls the presentation of a respective view of a virtual area and objects in the virtual area on a display monitor. The So3D engine typically interfaces with a graphical user interface driver and the HID devices to present the views of the virtual area and to allow the communicant to control the operation of the communications application 26.
[00132] In some embodiments, the So3D engine receives graphics rendering instructions from the area service. The So3D engine also may read a local
communicant avatar database that contains images needed for rendering the communicant's avatar in the virtual area. Based on this information, the So3D engine generates a visual representation (i.e., an image) of the virtual area and the objects in the virtual area from the point of view (position and orientation) of the communicant's avatar in the virtual area. The visual representation typically is passed to the graphics rendering components of the operating system, which drive the graphics rendering hardware to render the visual representation of the virtual area on the client network node.
[001331 ~^e communicant can control the presented view of the virtual area by inputting view control commands via a HID device (e.g., a computer mouse) The So3D engine updates the view of the virtual area in accordance with the view control commands. The So3D engine also updates the graphic representation of the virtual area on the display monitor in accordance with updated object position information received from the area service.
3. SYSTEM DATABASE AND STORAGE FACILITY
[00134] The system database and storage facility stores various kinds of information that is used by the platform. Exemplary information that typically is stored by the storage facility includes the presence database, the relationship database, an avatar database, a real user id (RUID) database, an art cache database, and an area application database. This information may be stored on a single network node or it may be distributed across multiple network nodes.
F. CLIENT NODE ARCHITECTURE
[00135] A communicant typically connects to the network 18 from a client network node. The client network node typicaily is implemented by a general-purpose computer system or a dedicated communications computer system (or ''console", such as a network-enabled video game console). The client network node executes communications processes that establish realtime data stream connections with other network nodes and typically executes visualization rendering processes that present a view of each virtual area entered by the communicant.
[00136] FIG. 19 shows an embodiment of a client network node that is implemented by a computer system 320. The computer system 320 includes a processing unit 322: a system memory 324, and a system bus 326 that couples the processing unit 322 to the various components of the computer system 320. The processing unit 322 may include one or more data processors, each of which may be in the form of any one of various commercially available computer processors. The system memory 324 includes one or more computer-readable media that typically are associated with a software application addressing space that defines the addresses that are available to software applications. The system memory 324 may include a read only memory (ROM) that stores a basic input/output system (BIOS) that contains startup routines for the computer system 320, and a random access memory (RAM). The system bus 326 may be a memory bus, a peripheral bus or a local bus, and may be compatible with any of a variety of bus protocols, including PCi, VESA, MicroChannel. ISA, and EISA. The computer system 320 also includes a persistent storage memory 328 {e.g. , a hard drive, a floppy drive, a CD ROM drive, magnetic tape drives, flash memory devices, and digital video disks) that is connected to the system bus 328 and contains one or more computer-readable media disks that provide non-volatile or persistent storage for data, data structures and computer-executable instructions.
[00137) A communicant may interact (e.g., input commands or data) with the computer system 320 using one or more input devices 330 (e.g. one or more
keyboards, computer mice, microphones, cameras, joysticks, physical motion sensors such Wii input devices, and touch pads), information may be presented through a graphical user interface (GUI) that is presented to the communicant on a display monitor 332, which is controlled by a display controller 334. The computer system 320 also may include other input/output hardware (e.g., peripheral output devices, such as speakers and a printer). The computer system 320 connects to other network nodes through a network adapter 336 (also referred to as a "network interface card" or NSC). [00138] A number of program modules may be stored in the system memory 324, including application programming interfaces 338 (APis), an operating system (OS) 340 (e.g., the Windows XP© operating system available from Microsoft Corporation of Redmond, Washington U.S.A.), the communications application 2S1 drivers 342 (e.g., a GUi driver), network transport protocols 344, and data 348 (e.g., input data, output data, program data, a registry, and configuration settings).
G. SERVER NODE ARCHITECTURE
[001391 *n some embodiments, the one or more server network nodes of the virtual environment creator 16 are implemented by respective general-purpose computer systems of the same type as the client network node 120, except that each server network node typically includes one or more server software applications.
[0014OJ in other embodiments, the one or more server network nodes of the virtual environment creator 16 are implemented by respective network devices that perform edge services (e.g., routing and switching).
H. EXEMPLARY COMMUNiCATiON SESSION
[00141 ] Referring back to FiG. 17, during a communication session, each of the client network nodes generates a respective set of realtime data streams (e.g., motion data streams, audio data streams, chat data streams, file transfer data streams, and video data streams). For example, each communicant manipulates one or more input devices (e.g., the computer mouse 52 and the keyboard 54) that generate motion data streams, which control the movement of his or her avatar in the virtual area 66, Sn addition, the communicant's voice and other sounds that are generated locally in the vicinity of the computer system 48 are captured by the microphone 60. The microphone 60 generates audio signals that are converted into realtime audio streams. Respective copies of the audio streams are transmitted to the other network nodes that are represented by avatars in the virtual area 66. The sounds that are generated locally at these other network nodes are converted into realtime audio signals and transmitted to the computer system 48. The computer system 48 converts the audio streams generated by the other network nodes into audio signals that are rendered by the speakers 56, 58. The motion data streams and audio streams may be transmitted from each of the communicant nodes to the other client network nodes either directly or indirectly. Sn some stream handling topologies, each of the client network nodes receives copies of the realtime data streams that are transmitted by the other client network nodes, in other stream handling topologies, one or more of the client network nodes receives one or more stream mixes that are derived from realtime data streams that are sourced (or originated) from other ones of the network nodes.
[00142| *n some embodiments, the area service maintains global state information that includes a current specification of the virtual area, a current register of the objects that are in the virtual area, and a list of any stream mixes that currently are being generated by the network node hosting the area service. The objects register typically includes for each object in the virtual area a respective object identifier (e.g., a label that uniquely identifies the object), a connection handle (e.g., a URI, such as an IP address) that enables a network connection to be established with a network node that is associated with the object, and interface data that identifies the realtime data sources and sinks that are associated with the object (e.g., the sources and sinks of the network node that is associated with the object). The objects register aiso typically includes one or more optional role identifiers for each object; the role identifiers may be assigned explicitly to the objects by either the communicants or the area service, or may be inferred from other attributes of the objects or the user. In some embodiments, the objects register also includes the current position of each of the objects in the virtual area as determined by the area service from an analysis of the realtime motion data streams received from the network nodes associated with objects in the virtual area Sn this regard, the area service receives realtime motion data streams from the network nodes associated with objects in the virtual area, tracks the communicants' avatars and other objects that enter, leave, and move around in the virtual area based on the motion data. The area service updates the objects register in accordance with the current locations of the tracked objects.
[00143] Sn the process of administering realtime data stream connections with other network nodes, the area service maintains for each of the client network nodes a set of configuration data: including interface data, a zone list, and the positions of the objects that currently are in the virtual area. The interface data includes for each object associated with each of the client network nodes a respective list of all the sources and sinks of realtime data stream types that are associated with the object The zone list is a register of all the zones in the virtual area that currently are occupied by the avatar associated with the corresponding client network node. When a communicant first enters a virtual area, the area service typically initializes the current object positions database with position initialization information. Thereafter, the area service updates the current object positions database with the current positions of the objects in the virtual area as determined from an analysts of the reaitsmβ motion data streams received from the other client network nodes sharing the virtual area.
I. INTERFACING WfTH A SPATIAL VIRTUAL COMMUNfCATiON
ENVIRONMENT
[00144] in addition to the local Human interface Device (HID) and audio playback devices, the So3D graphical display, avatar, and physics engine, and the system database and storage facility, the communications application 26 also includes a graphical navigation and interaction interface (referred to herein as a "seeker interface") that interfaces the user with the spatial virtual communication environment. The seeker interface includes navigation controls that enable the user to navigate the virtual environment and interaction controls that enable the user to control his or her interactions with other communicants in the virtual communication environment. The navigation and interaction controls typically are responsive to user selections that are made using any type of input device, including a computer mouse, a touch pad, a touch screen display, a keyboard, and a video game controller The seeker interface is an application that operates on each client network node. The seeker interface is a small, lightweight interface that a user can keep up and running aii the time on his or her desktop. The seeker interface allows the user to launch vtrtuaS area applications and provides the user with immediate access to realtime contacts and realtime collaborative places (or areas) The seeker interface is integrated with realtime communications applications and/or realtime communications components of the underlying operating system such that the seeker interface can initiate and receive realtime communications with other network nodes. A virtual area is integrated with the user's desktop through the seeker interface such that the user can upload files into the virtual environment created by the virtual environment creator 16, use files stored in association with the virtual area using the native client software applications independently of the virtual environment while still present in a virtual area, and more generally treat presence and position within a virtual area as an aspect of their operating environment analogous to other operating system functions rather than just one of several applications. [00145] Additional details regarding the construction and operation of embodiments of the seeker interface are described in co-pending U.S. Patent
Application No. 12/354,709, filed January 15, 2009.
[00148| Any of the embodiments of the spatial interfaces that are described herein may be integrated into the seeker interface in order to provide a context for depicting the current communication of the communicants involved in realtime networked communications. Embodiments of these spatial interfaces also provide a context for organizing the presentation of various interface elements that are used by communicants to participate in realtime networked communications, as described above.
V. CONCLUSION
[00147] The embodiments that are described herein provide improved systems and methods for visualizing reaϊtime network communications, in particular, these embodiments apply a spatial metaphor on top of realtime networked communications. The spatial metaphor provides a context for depicting the current communication state of the communicants involved in realtime networked communications. The spatial metaphor also provides a context for organizing the presentation of various interface elements that are used by communicants to participate in realtime networked communications.
[001481 Other embodiments are within the scope of the claims.

Claims

1. A computer-implemented method, comprising:
establishing a current realtime communication session between communicants operating on respective network nodes (12, 14);
on a display (332), displaying a spatial visualization (70, 92, 188, 220) of the current realtime communication session, wherein the spatial visualization (70, 92 188, 220) comprises a graphical representation (46, 48) of each of the communicants in spatial relation to a graphical representation (50) of a virtual area; and
during the current communication session, depicting visual cues in the spatial visualization (70. 92, 188, 220) that show current communication states of the communicants, wherein each of the communication states corresponds to a state of a respective communication channel over which a respective one of the communicants is configured to communicate. 2. The method of claim 1 , further comprising, during the current
communication session, presenting on the display (332) a log (114) of event
descriptions describing respective events involving interactions of the communicants in the virtual area in contextual association with elements of the spatial visualization (70, 92: 188: 220) of the current realtime communication session
3. The method of claim 2, wherein the log (1 14) of event descriptions and the graphical representation of the virtual area are displayed in a single graphical user interface window (92). 4, The method of claim 2, wherein the log (114) of event descriptions comprises at least one of: text of a chat conversation between the communicants in the virtual area; a description of a data file shared by a respective one of the communicants in the virtual area; and a description of an application shared by a respective one of the communicants in the virtual area.
5, The method of claim 2, wherein the presenting comprises visually associating the event descriptions in the log (114) with respective ones of the graphical representations of the communicants involved in the events described by the respective event descriptions.
8. The method of claim 5, wherein the visually associating comprises associating with each of the event descriptions a respective iabe! that has a respective visual appearance that matches a visual element of the graphical representation (46, 48) of the communicant involved in the event described by the respective event description. 7. The method of claim 2, further comprising storing the iog (114) of event descriptions in one or more database records that are indexed by an identifier of the virtual area.
8. The method of claim 1 , wherein the displaying comprises displaying in the virtual area one or more props (104-108) each representing a respective communication channel for realtime communications between the communicants during the
communication session,
9. The method of claim 8, wherein the displaying comprises displaying a communicant-selectable table prop (108) in the virtual area, and further comprising initiating a file share session between the communicants in response to selection of the table prop (108) by one of the communicants.
10. The method of claim 8, wherein the displaying comprises displaying a communicant-selectable viewscreen prop (104, 106) in the virtual area, and further comprising initiating an application sharing session between the communicants in response to selection of the viewscreen prop (104, 106) by one of the communicants,
11. The method of claim 8, further comprising changing a spatial property of the graphical representation of a respective one of the communicants in relation to a respective one of the props {104-108} in response to selection of the respective prop by the respective communicant.
12. The method of claim 11 , wherein the changing comprises depicting the graphical representation (46, 48) of the respective communicant adjacent the selected prop (104-108). 13. The method of cϊaim 11 , wherein the changing comprises reorienting the graphicaS representation (46, 48) of the respective communicant to face the selected prop (104-108).
14. The method of claim 11 , wherein the changing comprises changing the graphical representation of the respective communicant (46, 48),
15. The method of claim 1 , wherein the establishing comprises establishing during the current communication session a realtime instant messaging communication channel between the communicants.
16. The method of claim 15, wherein the displaying comprises displaying in association with the graphical representation of the virtual area a current chat Sog (114) of a current chat conversation between the communicants occurring during the current communication session.
17. The method of claim 16, wherein the depicting comprises dynamically modulating the graphical representation (46, 48} of a given one of the communicants in response to receipt of a respective realtime chat stream from the given communicant over the realtime instant messaging communication channel such that the current communication state of the given communicant is reflected in the dynamic modulation of the graphicaS representation (46, 48) of the given communicant.
18. The method of claim 16, wherein the displaying comprises displaying in association with the current chat log (114) a respective prior chat log (114) of a prior chat conversation that occurred during a prior communication session between the communicants in the virtual area
19. The method of claim 1 , wherein the displaying comprises dispiaying a graphical representation of a file sharing prop (108, 230) in the virtual area, and further comprising; in response to seiection of the file sharing prop by a respective one of the communicants, depicting the graphical representation of the respective communicant adjacent the fiie sharing prop (108, 230), and initiating a realtime file sharing session in the virtual area.
20. The method of claim 19, further comprising storing a data file shared by the respective communicant during the realtime file sharing session in a data storage device with an index comprising an identifier of the virtual area, and wherein the displaying comprises displaying on the file sharing prop (230) a communicant-selectable graphical representation (231 ) of the data file.
21. The method of claim 20, further comprising initiating a download of the data file to the network node from which a given one of the communicants is operating in response to selection of the graphical representation (231 ) of the file by the given communicant,
22. The method of claim 1 , wherein the displaying comprises displaying a graphical representation of an application sharing prop (226, 228) in the virtual area, and further comprising: in response to selection of the application sharing prop by a respective one of the communicants, depicting the graphical representation of the respective communicant adjacent the application sharing prop, and initiating a realtime application sharing session in the virtual area.
23. The method of claim 22, further comprising sharing screen shots from the network node from which the respective communicant is operating with one or more of the other communicants during the realtime application sharing session, and wherein the displaying comprises displaying a graphical indication that an application that is being shared in connection with the application sharing prop (226, 228).
24. The method of claim 22, wherein the displaying comprises displaying a first graphical representation of the application sharing prop (226, 228} during periods of application sharing between the communicants in the virtual area and dispiaying a second graphical representation of the appiication sharing prop different from the first graphical representation during periods free of application sharing between the communicants.
25. The method of claim 1. wherein in response to a command from a given one of the communicants to activate an audio sink communication channel, the establishing comprises establishing a realtime audio communication channel between the given communicant and one or more of the other communicants configured as audio sources, and the depicting comprises modifying the graphical representation of the given communicant (46, 48) to show that the given communicant is configured as an audio sink.
26. The method of claim 1 , wherein in response to a command from a given one of the communicants to activate an audio source communication channel, the establishing comprises establishing a reaitime audio communication channel between the given communicant and one or more of the other communicants configured as audio sinks, and the depicting comprises modifying the graphical representation (46, 48) of the given communicant to show that the given communicant is configured as an audio source.
27. The method of claim 1 , wherein the displaying comprises displaying a static view of the graphical representation of the virtual area throughout the current communication session, and the communicants are unable to navigate the graphical representations (46. 48) of the communicants outside the static view of the virtual area.
28. The method of claim 1 , wherein in response to receipt of a command from a first one of the communicants to initiate a private communication with a second one of the communicants; the establishing comprises establishing the current realtime communication session between the first and second communicants, and the displaying comprises displaying the graphical representations {45, 48) of the first and second communicants in spatial relation to a graphical representation (50) of a virtual area that is indexed by identifiers of the first and second communicants.
29. The method of claim 1 , further comprising determining an end state of a prior realtime communication session between the communicants from data that is indexed by an identifier of the virtual area and describes events that occurred during a prior communication session between the communicants: and wherein the displaying comprises displaying the graphical representation (50) of a virtual area in a state that corresponds to the determined end state of the prior communication session between the communicants. 30. Apparatus, comprising:
a computer-readable medium (324, 328} storing computer-readable instructions; and
a data processor (322) coupled to the computer-readable medium, operable to execute the instructions, and based at least in part on the execution of the instructions operable to perform operations comprising
establishing a current realtime communication session between
communicants operating on respective network nodes (12; 14), on a display (332), displaying a spatial visualization (70, 92, 188, 220) of the current realtime communication session, wherein the spatial visuaiization (70, 92, 188, 220) comprises a graphical
representation (48, 48) of each of the communicants in spatial relation to a graphical representation (50) of a virtual area; and during the current communication session, depicting visual cues in the spatial visuaiization (70. 92, 188, 220) that show current communication states of the communicants, wherein each of the communication states corresponds to a state of a respective communication channel over which a respective one of the communicants is configured to communicate. 31 , At least one computer-readable medium (324, 328) having computer- readable program code embodied therein, the computer-readable program code adapted to be executed by a computer (320) to implement a method comprising: establishing a current realtime communication session between communicants operating on respective network nodes (12, 14);
on a display (332), displaying a spatial visualization (70, 92, 188, 220) of the current realtime communication session, wherein the spatia! visualization (70, 92, 188, 220) comprises a graphical representation (48, 48) of each of the communicants in spatial relation to a graphical representation (50) of a virtual area: and
during the current communication session, depicting visual cues in the spatial visualization (70, 92, 188, 220) that show current communication states of the communicants, wherein each of the communication states corresponds to a state of a respective communication channel over which a respective one of the communicants is configured to communicate.
32. A computer-implemented method, comprising:
establishing a current realtime communication session between communicants operating on respective network nodes (12, 14);
on a display (332), displaying a spatial visualization (70, 92, 18S1 220) of the current realtime communication session, wherein the spatia! visualization (70, 92, 188, 220) comprises a graphical representation (46, 48) of each of the communicants in spatial relation to a graphical representation (50) of a virtual area; and
during the current communication session, on the display (332) presenting a log
(1 14) of event descriptions describing respective events involving interactions of the communicants in the virtual area, wherein the event descriptions are presented in contextual association with elements of the spatia! visualization (70, 92, 188, 220) of the current realtime communication session.
33. The method of claim 32, wherein the presenting comprises depicting a visual association between respective ones of the event descriptions in the log (114) and elements of the spatial visualization (70, 92. 188, 220} of the current realtime communication session.
34. The method of claim 33, wherein the depicting comprises depicting a visual association between respective ones of the event descriptions in the log (114) with respective ones of the graphical representations of the communicants involved in the events described by the respective event descriptions.
35. The method of claim 34, wherein the depicting comprises associating with each of one or more of the event descriptions a respective label that has a respective visual appearance that matches a visual element of the graphical representation (46, 48) of the communicant involved in the event described by the event description.
36. The method of claim 32, wherein in response to an entry of a respective one of the communicants into the virtual area, the displaying comprises adding the graphical representation of the respective communicant to the spatial visualization (70, 92, 188, 220), and the presenting comprises presenting a respective one of the event descriptions describing the entry of the respective communicant into the virtual area. 37. The method of claim 32, wherein in response to a departure of a respective one of the communicants from the virtual area, the displaying comprises removing the graphical representation (46, 48) of the respective communicant from the spatial visualization (70, 92, 188, 220), and the presenting comprises presenting a respective one of the event descriptions describing the departure of the respective communicant from the virtual area.
38. The method of claim 32, wherein in response to a sharing of a data file by a respective one of the communicants with other ones of the communicants, the displaying comprises displaying a communicant-selectable graphical representation (231 } of the data file in spatial relation to the graphical representation of the virtual area, and the presenting comprises presenting a respective one of the event descriptions describing the sharing of the data file by the respective communicant.
39. The method of claim 32, wherein in response to a sharing of an application by a respective one of the communicants with other ones of the communicants, the displaying comprises displaying a graphical indication of the sharing of the application in spatial relation to the graphical representation of the virtual area, and the presenting comprises presenting a respective one of the event descriptions describing the sharing of the application by the respective communicant.
40. Apparatus, comprising:
a computer-readable medium (324, 328) storing computer-readable instructions; and
a data processor (322) coupled to the computer-readable medium, operable to execute the instructions, and based at least in part on the execution of the instructions operable to perform operations comprising
establishing a current realtime communication session between
communicants operating on respective network nodes (12, 14). on a display (332), displaying a spatial visualization (70, 92, 188, 220) of the current realtime communication session, wherein the spatial visualization (70, 92, 188, 220) comprises a graphical
representation (46: 48} of each of the communicants in spatial relation to a graphical representation (50) of a virtual area, and during the current communication session, on the display (332) presenting a log (1 14) of event descriptions describing respective events involving interactions of the communicants in the virtual area, wherein the event descriptions are presented in contextual association with elements of the spatiai visualization (70, 92, 188, 220) of the current realtime communication session.
41 . At least one computer-readable medium (324, 328) having computer- readable program code embodied therein, the computer-readable program code adapted to be executed by a computer (320) to implement a method comprising:
establishing a current realtime communication session between communicants operating on respective network nodes (12, 14);
on a display (332), displaying a spatial visualization (70, 92, 188, 220) of the current realtime communication session, wherein the spatial visualization (70, 92, 188,
220) comprises a graphical representation (48, 48) of each of the communicants in spatial relation to a graphical representation (50) of a virtual area: anά during the current communication session, on the display (332) presenting a log {114) of event descriptions describing respective events involving interactions of the communicants in the virtual area, wherein the event descriptions are presented in contextua! association with elements of the spatia! visualization (70, 92, 188, 220) of the current realtime communication session.
42, A computer-implemented method, comprising in response to receipt of a command from a first communicant operating on a first network node (12) to initiate a private communication with a second communicant operating on a second network node (14):
establishing a current realtime communication session between the first and second network nodes (12, 14);
identifying a private virtual area associated with the first and second
communicants;
retrieving context configuration data associated with the private virtual area and generated in response to interactions of the first and second communicants in the private virtual area; and
on a display (332), displaying a spatial visualization (188, 220) of the current realtime communication session, wherein the spatial visualization (188, 220) comprises graphical representations of the first and second communicants in spatial relation to a graphical representation of the virtual area configured in accordance with the context configuration data.
43, The method of claim 42, further comprising during the current realtime communication session, generating a log (114) of event descriptions describing respective events involving interactions of the first and second communicants in the virtual area.
44, The method of claim 43, further comprising during the current realtime communication session, storing the event descriptions in a data storage device with an index comprising an identifier of the virtual area
45. The method of claim 44 s wherein the log (114) of event descriptions comprises at least one of. text of a chat conversation between the first and second communicants in the virtual area; a description of a data file shared by a respective one of the first and second communicants in the virtual area; and a description of an application shared by a respective one of the first and second communicants in the virtual area,
46. The method of claim 43, further comprising during the current realtime communication session, presenting the log (114) of event descriptions on the display (332).
47. The method of claim 46, wherein the presenting comprises presenting the log (114) of event descriptions in contextual association with elements of the spatial visualization (188, 220} of the current realtime communication session.
48. The method of claim 46, wherein the retrieving comprises retrieving context configuration data comprising a log (114) of event descriptions describing respective events involving interactions of the first and second communicants in the virtual area during one or more prior communication sessions before the current communication session, and the presenting comprises presenting the log (114) of event descriptions generated during the current realtime communication session together with the retrieved context configuration data comprising the log (114) of event descriptions.
49. The method of claim 42, wherein the retrieving comprises retrieving context configuration data comprising a description of an end state of a prior realtime communication session between the communicants, and the displaying comprises displaying the graphical representation (50) of a virtual area in a state that corresponds to the end state of the prior communication session between the communicants. 50. Apparatus, comprising.
a computer-readable medium (324; 328) storing computer-readable instructions; and a data processor (322) coupled to the computer-readable medium, operable to execute the instructions, and based at least in part on the execution of the instructions operable to perform operations comprising in response to receipt of a command from a first communicant operating on a first network node to initiate a private communication with a second communicant operating on a second network node,
establishing a current realtime communication session between the first and second network nodes (12, 14),
identifying a private virtual area associated with the first and second
communicants,
retrieving context configuration data associated with the private virtual area and generated in response to interactions of the first and second communicants in the private virtual area, and
on a display (332), displaying a spatial visualization (188. 220) of the
current realtime communication session, wherein the spatial visualization (188, 220) comprises graphical representations of the first and second communicants in spatial relation to a graphical representation of the virtual area configured in accordance with the context configuration data. 51 . At least one computer-readable medium (324, 328) having computer- readable program code embodied therein, the computer-readable program code adapted to be executed by a computer (320) to implement a method comprising:
in response to receipt of a command from a first communicant operating on a first network node to initiate a private communication with a second communicant operating on a second network node,
establishing a current realtime communication session between the first and second network nodes (12, 14),
identifying a private virtual area associated with the first and second
communicants,
retrieving context configuration data associated with the private virtual area and generated in response to interactions of the first and second communicants in the private virtual area, and splay (332), displaying a spatial visualization (188, 220) of the current realtime communication session, wherein the spatial visualization (188, 220} comprises graphical representations of the first and second communicants in spatial relation to a graphical representation of the virtual area configured in accordance with the context configuration data.
PCT/US2010/042119 2009-07-27 2010-07-15 Spatial interfaces for realtime networked communications WO2011016967A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN2010800346974A CN102483819A (en) 2009-07-27 2010-07-15 Spatial Interfaces For Realtime Networked Communications
EP10806818A EP2460138A2 (en) 2009-07-27 2010-07-15 Spatial interfaces for realtime networked communications
IL217290A IL217290A0 (en) 2009-07-27 2011-12-29 Spatial interfaces for realtime networked communications

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/509,658 US20090288007A1 (en) 2008-04-05 2009-07-27 Spatial interfaces for realtime networked communications
US12/509,658 2009-07-27

Publications (2)

Publication Number Publication Date
WO2011016967A2 true WO2011016967A2 (en) 2011-02-10
WO2011016967A3 WO2011016967A3 (en) 2011-04-14

Family

ID=43544836

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/042119 WO2011016967A2 (en) 2009-07-27 2010-07-15 Spatial interfaces for realtime networked communications

Country Status (6)

Country Link
US (1) US20090288007A1 (en)
EP (1) EP2460138A2 (en)
KR (1) KR20120050980A (en)
CN (1) CN102483819A (en)
IL (1) IL217290A0 (en)
WO (1) WO2011016967A2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013181026A1 (en) * 2012-06-02 2013-12-05 Social Communications Company Interfacing with a spatial virtual communications environment
CN107688418A (en) * 2017-05-05 2018-02-13 平安科技(深圳)有限公司 Network indicates the methods of exhibiting and system of control

Families Citing this family (70)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4697953B2 (en) * 2005-09-12 2011-06-08 キヤノン株式会社 Image display device and image display method
US9357025B2 (en) 2007-10-24 2016-05-31 Social Communications Company Virtual area based telephony communications
US8397168B2 (en) * 2008-04-05 2013-03-12 Social Communications Company Interfacing with a spatial virtual communication environment
US8407605B2 (en) 2009-04-03 2013-03-26 Social Communications Company Application sharing
US7769806B2 (en) 2007-10-24 2010-08-03 Social Communications Company Automated real-time data stream switching in a shared virtual area communication environment
US9009603B2 (en) 2007-10-24 2015-04-14 Social Communications Company Web browser interface for spatial communication environments
US8930472B2 (en) 2007-10-24 2015-01-06 Social Communications Company Promoting communicant interactions in a network communications environment
US9514444B2 (en) 2009-01-15 2016-12-06 Sococo, Inc. Encapsulating virtual area based communicant assemblies
CN102084354A (en) 2008-04-05 2011-06-01 社会传播公司 Shared virtual area communication environment based apparatus and methods
US8405702B1 (en) 2008-11-24 2013-03-26 Shindig, Inc. Multiparty communications systems and methods that utilize multiple modes of communication
US9401937B1 (en) 2008-11-24 2016-07-26 Shindig, Inc. Systems and methods for facilitating communications amongst multiple users
KR20110106869A (en) 2008-12-05 2011-09-29 소우셜 커뮤니케이션즈 컴퍼니 Pervasive realtime framework
US9853922B2 (en) * 2012-02-24 2017-12-26 Sococo, Inc. Virtual area communications
US9319357B2 (en) 2009-01-15 2016-04-19 Social Communications Company Context based virtual area creation
US9065874B2 (en) 2009-01-15 2015-06-23 Social Communications Company Persistent network resource and virtual area associations for realtime collaboration
US20130283169A1 (en) 2012-04-24 2013-10-24 Social Communications Company Voice-based virtual area navigation
US9288242B2 (en) * 2009-01-15 2016-03-15 Social Communications Company Bridging physical and virtual spaces
JP5669418B2 (en) * 2009-03-30 2015-02-12 アバイア インク. A system and method for managing incoming requests that require a communication session using a graphical connection display.
US9344745B2 (en) 2009-04-01 2016-05-17 Shindig, Inc. Group portraits composed using video chat systems
US8779265B1 (en) 2009-04-24 2014-07-15 Shindig, Inc. Networks of portable electronic devices that collectively generate sound
US8244754B2 (en) 2010-02-01 2012-08-14 International Business Machines Corporation System and method for object searching in virtual worlds
US8639756B2 (en) * 2010-02-19 2014-01-28 Nokia Corporation Method and apparatus for generating a relevant social graph
US20110225515A1 (en) * 2010-03-10 2011-09-15 Oddmobb, Inc. Sharing emotional reactions to social media
US20110225517A1 (en) * 2010-03-10 2011-09-15 Oddmobb, Inc Pointer tools for a virtual social venue
US20110225498A1 (en) * 2010-03-10 2011-09-15 Oddmobb, Inc. Personalized avatars in a virtual social venue
US8572177B2 (en) * 2010-03-10 2013-10-29 Xmobb, Inc. 3D social platform for sharing videos and webpages
US20110239136A1 (en) * 2010-03-10 2011-09-29 Oddmobb, Inc. Instantiating widgets into a virtual social venue
US20110225039A1 (en) * 2010-03-10 2011-09-15 Oddmobb, Inc. Virtual social venue feeding multiple video streams
US20110225518A1 (en) * 2010-03-10 2011-09-15 Oddmobb, Inc. Friends toolbar for a virtual social venue
US8667402B2 (en) * 2010-03-10 2014-03-04 Onset Vi, L.P. Visualizing communications within a social setting
US20110225519A1 (en) * 2010-03-10 2011-09-15 Oddmobb, Inc. Social media platform for simulating a live experience
US20110225516A1 (en) * 2010-03-10 2011-09-15 Oddmobb, Inc. Instantiating browser media into a virtual social venue
US9955209B2 (en) 2010-04-14 2018-04-24 Alcatel-Lucent Usa Inc. Immersive viewer, a method of providing scenes on a display and an immersive viewing system
WO2011136794A1 (en) * 2010-04-30 2011-11-03 America Teleconferencing Services, Ltd Record and playback in a conference
US20110276901A1 (en) * 2010-05-04 2011-11-10 Qwest Communications International Inc. Family chat
US9356790B2 (en) 2010-05-04 2016-05-31 Qwest Communications International Inc. Multi-user integrated task list
US9501802B2 (en) 2010-05-04 2016-11-22 Qwest Communications International Inc. Conversation capture
US9003306B2 (en) 2010-05-04 2015-04-07 Qwest Communications International Inc. Doodle-in-chat-context
US8819566B2 (en) 2010-05-04 2014-08-26 Qwest Communications International Inc. Integrated multi-modal chat
US9559869B2 (en) 2010-05-04 2017-01-31 Qwest Communications International Inc. Video call handling
KR20120055371A (en) * 2010-11-23 2012-05-31 한국전자통신연구원 Method and apparatus for protecting digital contents
US9276972B2 (en) 2010-12-14 2016-03-01 Microsoft Technology Licensing, Llc Real-time media optimization over remoted sessions
US20120216129A1 (en) * 2011-02-17 2012-08-23 Ng Hock M Method and apparatus for providing an immersive meeting experience for remote meeting participants
WO2012135231A2 (en) 2011-04-01 2012-10-04 Social Communications Company Creating virtual areas for realtime communications
US9053750B2 (en) 2011-06-17 2015-06-09 At&T Intellectual Property I, L.P. Speaker association with a visual representation of spoken content
US8719031B2 (en) 2011-06-17 2014-05-06 At&T Intellectual Property I, L.P. Dynamic access to external media content based on speaker content
US9411506B1 (en) * 2011-06-28 2016-08-09 Google Inc. Providing additional functionality for a group messaging application
IL214855A0 (en) * 2011-08-28 2011-10-31 Arnon Joseph A method and device for carrying out a computerized group session
CN102413140A (en) * 2011-11-30 2012-04-11 江苏奇异点网络有限公司 Network teaching method for supporting speech interaction
US9563902B2 (en) * 2012-04-11 2017-02-07 Myriata, Inc. System and method for transporting a virtual avatar within multiple virtual environments
KR101426994B1 (en) * 2012-06-29 2014-08-05 인텔렉추얼디스커버리 주식회사 Apparatus and method of sharing data using topology of mobile portable devices
GB2506102A (en) * 2012-07-25 2014-03-26 Nowhere Digital Ltd Meeting management system
US20140173466A1 (en) * 2012-12-14 2014-06-19 Microsoft Corporation Transitions within views of conversation environments
US9679331B2 (en) * 2013-10-10 2017-06-13 Shindig, Inc. Systems and methods for dynamically controlling visual effects associated with online presentations
US20150120840A1 (en) * 2013-10-29 2015-04-30 International Business Machines Corporation Resource referencing in a collaboration application system and method
US10271010B2 (en) 2013-10-31 2019-04-23 Shindig, Inc. Systems and methods for controlling the display of content
CN103744592B (en) * 2013-12-26 2017-11-21 华为技术有限公司 The method and terminal of a kind of information processing
US9733333B2 (en) 2014-05-08 2017-08-15 Shindig, Inc. Systems and methods for monitoring participant attentiveness within events and group assortments
WO2016164929A1 (en) * 2015-04-10 2016-10-13 Juggle, Inc. System and method for visually facilitated contact interaction management
US10235366B2 (en) * 2016-08-16 2019-03-19 Microsoft Technology Licensing, Llc Activity gallery view in communication platforms
US10133916B2 (en) 2016-09-07 2018-11-20 Steven M. Gottlieb Image and identity validation in video chat events
US10452410B2 (en) 2016-10-25 2019-10-22 International Business Machines Corporation Context aware user interface
CN106453602B (en) * 2016-10-28 2019-10-29 深圳多哚新技术有限责任公司 A kind of data processing method and device based on VR glasses
CN108573519B (en) 2017-03-14 2022-02-11 阿里巴巴集团控股有限公司 Method and device for generating three-dimensional graphic file and displaying three-dimensional graphic at client
US11381413B2 (en) * 2020-01-08 2022-07-05 Disney Enterprises, Inc. Audio-orientated immersive experience of an event
US20220070239A1 (en) * 2020-08-28 2022-03-03 Tmrw Foundation Ip S. À R.L. System and method to provision cloud computing-based virtual computing resources within a virtual environment
US20220070235A1 (en) 2020-08-28 2022-03-03 Tmrw Foundation Ip S.Àr.L. System and method enabling interactions in virtual environments with virtual presence
EP4334802A1 (en) * 2021-05-07 2024-03-13 Telefonaktiebolaget LM Ericsson (publ) Method and arrangements for graphically visualizing data transfer in a 3d virtual environment
US11831814B2 (en) * 2021-09-03 2023-11-28 Meta Platforms Technologies, Llc Parallel video call and artificial reality spaces
US11921970B1 (en) 2021-10-11 2024-03-05 Meta Platforms Technologies, Llc Coordinating virtual interactions with a mini-map

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2000070557A2 (en) * 1999-05-14 2000-11-23 Graphic Gems Method and apparatus for registering lots in a shared virtual world
US6396509B1 (en) * 1998-02-21 2002-05-28 Koninklijke Philips Electronics N.V. Attention-based interaction in a virtual environment
US6772195B1 (en) * 1999-10-29 2004-08-03 Electronic Arts, Inc. Chat clusters for a virtual world application
US7181690B1 (en) * 1995-11-13 2007-02-20 Worlds. Com Inc. System and method for enabling users to interact in a virtual space
US20070233785A1 (en) * 2006-03-30 2007-10-04 International Business Machines Corporation Communicating using collaboration spaces
EP1964597A1 (en) * 2007-03-01 2008-09-03 Sony Computer Entertainment Europe Ltd. Apparatus and method of modifying an online environment

Family Cites Families (96)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5745711A (en) * 1991-10-23 1998-04-28 Hitachi, Ltd. Display control method and apparatus for an electronic conference
US5471318A (en) * 1993-04-22 1995-11-28 At&T Corp. Multimedia communications network
US5689641A (en) * 1993-10-01 1997-11-18 Vicor, Inc. Multimedia collaboration system arrangement for routing compressed AV signal through a participant site without decompressing the AV signal
US5491743A (en) * 1994-05-24 1996-02-13 International Business Machines Corporation Virtual conference system and terminal apparatus therefor
US5627978A (en) * 1994-12-16 1997-05-06 Lucent Technologies Inc. Graphical user interface for multimedia call set-up and call handling in a virtual conference on a desktop computer conferencing system
US5793365A (en) * 1996-01-02 1998-08-11 Sun Microsystems, Inc. System and method providing a computer user interface enabling access to distributed workgroup members
US6275490B1 (en) * 1996-08-21 2001-08-14 Netspeak Corporation Method and apparatus for establishing communications from browser application
US6862625B1 (en) * 1996-09-27 2005-03-01 Avaya Technology Corp. Method and apparatus for real time network communication
US5764916A (en) * 1996-09-27 1998-06-09 Ichat, Inc. Method and apparatus for real time communication over a computer network
US6057856A (en) * 1996-09-30 2000-05-02 Sony Corporation 3D virtual reality multi-user interaction with superimposed positional information display for each user
US6785708B1 (en) * 1996-10-30 2004-08-31 Avaya Inc. Method and apparatus for synchronizing browse and chat functions on a computer network
US7263526B1 (en) * 1996-10-30 2007-08-28 Avaya Technology Corp. Method and apparatus for embedding chat functions in a web page
US6119166A (en) * 1997-03-28 2000-09-12 International Business Machines Corporation Controlling communications with local applications using a browser application
GB9721667D0 (en) * 1997-10-14 1997-12-10 Philips Electronics Nv Virtual environment navigation aid
US6380952B1 (en) * 1998-04-07 2002-04-30 International Business Machines Corporation System for continuous display and navigation in a virtual-reality world
US5999208A (en) * 1998-07-15 1999-12-07 Lucent Technologies Inc. System for implementing multiple simultaneous meetings in a virtual reality mixed media meeting room
US6119147A (en) * 1998-07-28 2000-09-12 Fuji Xerox Co., Ltd. Method and system for computer-mediated, multi-modal, asynchronous meetings in a virtual space
US6731314B1 (en) * 1998-08-17 2004-05-04 Muse Corporation Network-based three-dimensional multiple-user shared environment apparatus and method
US6388688B1 (en) * 1999-04-06 2002-05-14 Vergics Corporation Graph-based visual navigation through spatial environments
US6292188B1 (en) * 1999-07-28 2001-09-18 Alltrue Networks, Inc. System and method for navigating in a digital information environment
JP2001154966A (en) * 1999-11-29 2001-06-08 Sony Corp System and method for supporting virtual conversation being participation possible by users in shared virtual space constructed and provided on computer network and medium storing program
US6396609B1 (en) * 1999-12-20 2002-05-28 Chorum Technologies, Lp Dispersion compensation for optical systems
US20060184886A1 (en) * 1999-12-22 2006-08-17 Urbanpixel Inc. Spatial chat in a multiple browser environment
US6708172B1 (en) * 1999-12-22 2004-03-16 Urbanpixel, Inc. Community-based shared multiple browser environment
US7054928B2 (en) * 1999-12-23 2006-05-30 M.H. Segan Limited Partnership System for viewing content over a network and method therefor
US7689696B2 (en) * 2000-01-31 2010-03-30 Telecommunication Systems, Inc. System and method for re-directing requests from browsers for communications over non-IP based networks
US7392306B1 (en) * 2000-04-07 2008-06-24 Aol Llc Instant messaging client having an embedded browser
JP3345600B2 (en) * 2000-04-10 2002-11-18 コナミ株式会社 Game system and computer-readable storage medium
US6784901B1 (en) * 2000-05-09 2004-08-31 There Method, system and computer program product for the delivery of a chat message in a 3D multi-user environment
US6714222B1 (en) * 2000-06-21 2004-03-30 E2 Home Ab Graphical user interface for communications
US7036082B1 (en) * 2000-09-21 2006-04-25 Nortel Networks Limited Controlling communications through a virtual reality environment
US8234218B2 (en) * 2000-10-10 2012-07-31 AddnClick, Inc Method of inserting/overlaying markers, data packets and objects relative to viewable content and enabling live social networking, N-dimensional virtual environments and/or other value derivable from the content
US7168051B2 (en) * 2000-10-10 2007-01-23 Addnclick, Inc. System and method to configure and provide a network-enabled three-dimensional computing environment
US20020075305A1 (en) * 2000-12-18 2002-06-20 Beaton Brian F. Graphical user interface for a virtual team environment
US7925703B2 (en) * 2000-12-26 2011-04-12 Numedeon, Inc. Graphical interactive interface for immersive online communities
US20030043200A1 (en) * 2001-08-09 2003-03-06 Urbanpixel Inc Interactive multi-level mapping in a multiple browser environment
US7660875B2 (en) * 2001-08-31 2010-02-09 Sony Corporation Bidirectional remote communication via browser plug-in
US7644144B1 (en) * 2001-12-21 2010-01-05 Microsoft Corporation Methods, tools, and interfaces for the dynamic assignment of people to groups to enable enhanced communication and collaboration
AUPR989802A0 (en) * 2002-01-09 2002-01-31 Lake Technology Limited Interactive spatialized audiovisual system
US7058896B2 (en) * 2002-01-16 2006-06-06 Silicon Graphics, Inc. System, method and computer program product for intuitive interactive navigation control in virtual environments
US7051049B2 (en) * 2002-02-21 2006-05-23 International Business Machines Corporation Real-time chat and conference contact information manager
US7336779B2 (en) * 2002-03-15 2008-02-26 Avaya Technology Corp. Topical dynamic chat
US7016978B2 (en) * 2002-04-29 2006-03-21 Bellsouth Intellectual Property Corporation Instant messaging architecture and system for interoperability and presence management
GB2391420A (en) * 2002-07-25 2004-02-04 Samsung Electronics Co Ltd Method for serving audio and image communication in web browser using session initiation protocol
US7234117B2 (en) * 2002-08-28 2007-06-19 Microsoft Corporation System and method for shared integrated online social interaction
US7676542B2 (en) * 2002-12-02 2010-03-09 Sap Ag Establishing a collaboration environment
US7474741B2 (en) * 2003-01-20 2009-01-06 Avaya Inc. Messaging advise in presence-aware networks
US20040158610A1 (en) * 2003-02-10 2004-08-12 Davis Joel A. Client proxying for instant messaging
US20040179039A1 (en) * 2003-03-03 2004-09-16 Blattner Patrick D. Using avatars to communicate
CA2525939C (en) * 2003-05-16 2015-09-22 Picasa, Inc. Networked chat and media sharing systems and methods
US8347313B2 (en) * 2003-05-21 2013-01-01 Resilient Networks, Inc. Method and apparatus for automating organization of processes
US7503006B2 (en) * 2003-09-25 2009-03-10 Microsoft Corporation Visual indication of current voice speaker
US7813488B2 (en) * 2003-09-29 2010-10-12 Siemens Enterprise Communications, Inc. System and method for providing information regarding an identity's media availability
US8528003B2 (en) * 2003-10-27 2013-09-03 Yahoo! Inc. Communication among browser windows
US7734691B2 (en) * 2003-12-18 2010-06-08 International Business Machines Corporation Providing collaboration services to a wireless device
US7707503B2 (en) * 2003-12-22 2010-04-27 Palo Alto Research Center Incorporated Methods and systems for supporting presentation tools using zoomable user interface
US7852997B2 (en) * 2004-01-28 2010-12-14 Managed Inventions, Llc Internet telephony communications adapter for web browsers
US7707249B2 (en) * 2004-09-03 2010-04-27 Open Text Corporation Systems and methods for collaboration
US7342587B2 (en) * 2004-10-12 2008-03-11 Imvu, Inc. Computer-implemented system and method for home page customization and e-commerce support
US7734692B1 (en) * 2005-07-22 2010-06-08 Oracle America, Inc. Network collaboration system with private voice chat
US7925000B2 (en) * 2005-08-29 2011-04-12 Avaya Inc. Managing held telephone calls from a remote telecommunications terminal
CA2632676C (en) * 2005-12-09 2018-05-22 Ebuddy Holding B.V. Message history display system and method
US7489772B2 (en) * 2005-12-30 2009-02-10 Nokia Corporation Network entity, method and computer program product for effectuating a conference session
US8266214B2 (en) * 2006-01-24 2012-09-11 Simulat, Inc. System and method for collaborative web-based multimedia layered platform with recording and selective playback of content
US20070198645A1 (en) * 2006-02-21 2007-08-23 Yen-Fu Chen Method for providing in-context responses to instant messaging conversations
US20070214424A1 (en) * 2006-03-13 2007-09-13 International Business Machines Corporation Networked chat technique
US20070220111A1 (en) * 2006-03-15 2007-09-20 Andrew Lin Personal communications browser client for remote use in enterprise communications
US20080052373A1 (en) * 2006-05-01 2008-02-28 Sms.Ac Systems and methods for a community-based user interface
US7680098B2 (en) * 2006-07-20 2010-03-16 Avaya Inc. Determining group availability on different communication media
US8295206B2 (en) * 2006-07-20 2012-10-23 Avaya Inc. Rule-based system for determining user availability
US7680480B2 (en) * 2006-07-20 2010-03-16 Avaya Inc. Determining user availability based on a past event
US8726195B2 (en) * 2006-09-05 2014-05-13 Aol Inc. Enabling an IM user to navigate a virtual world
US7698660B2 (en) * 2006-11-13 2010-04-13 Microsoft Corporation Shared space for communicating information
US7765259B2 (en) * 2006-12-05 2010-07-27 Avaya Inc. System and method for aggregation of user conversations and visualizing personal communications map
US8706169B2 (en) * 2006-12-28 2014-04-22 Yahoo! Inc. Interface overlay
US8554868B2 (en) * 2007-01-05 2013-10-08 Yahoo! Inc. Simultaneous sharing communication interface
US20080263460A1 (en) * 2007-04-20 2008-10-23 Utbk, Inc. Methods and Systems to Connect People for Virtual Meeting in Virtual Reality
US7840668B1 (en) * 2007-05-24 2010-11-23 Avaya Inc. Method and apparatus for managing communication between participants in a virtual environment
US8209384B2 (en) * 2007-10-23 2012-06-26 Yahoo! Inc. Persistent group-based instant messaging
US7499926B1 (en) * 2007-11-16 2009-03-03 International Business Machines Corporation Maintaining and replicating chat histories
US20090222742A1 (en) * 2008-03-03 2009-09-03 Cisco Technology, Inc. Context sensitive collaboration environment
US8605863B1 (en) * 2008-03-18 2013-12-10 Avaya Inc. Method and apparatus for providing state indication on a telephone call
US8531447B2 (en) * 2008-04-03 2013-09-10 Cisco Technology, Inc. Reactive virtual environment
US8312380B2 (en) * 2008-04-04 2012-11-13 Yahoo! Inc. Local map chat
US20090307189A1 (en) * 2008-06-04 2009-12-10 Cisco Technology, Inc. Asynchronous workflow participation within an immersive collaboration environment
US8468253B2 (en) * 2008-12-02 2013-06-18 At&T Intellectual Property I, L.P. Method and apparatus for multimedia collaboration using a social network system
US20100162121A1 (en) * 2008-12-22 2010-06-24 Nortel Networks Limited Dynamic customization of a virtual world
US8542232B2 (en) * 2008-12-28 2013-09-24 Avaya Inc. Method and apparatus for monitoring user attention with a computer-generated virtual environment
US20100169796A1 (en) * 2008-12-28 2010-07-01 Nortel Networks Limited Visual Indication of Audio Context in a Computer-Generated Virtual Environment
US9064023B2 (en) * 2008-12-29 2015-06-23 Avaya Inc. Providing web content in the context of a virtual environment
US20100169799A1 (en) * 2008-12-30 2010-07-01 Nortel Networks Limited Method and Apparatus for Enabling Presentations to Large Numbers of Users in a Virtual Environment
US20100228560A1 (en) * 2009-03-04 2010-09-09 Avaya Inc. Predictive buddy list-reorganization based on call history information
US8499085B2 (en) * 2009-03-16 2013-07-30 Avaya, Inc. Advanced availability detection
US8386255B2 (en) * 2009-03-17 2013-02-26 Avaya Inc. Providing descriptions of visually presented information to video teleconference participants who are not video-enabled
US20100246570A1 (en) * 2009-03-24 2010-09-30 Avaya Inc. Communications session preparation method and apparatus
JP5669418B2 (en) * 2009-03-30 2015-02-12 アバイア インク. A system and method for managing incoming requests that require a communication session using a graphical connection display.

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7181690B1 (en) * 1995-11-13 2007-02-20 Worlds. Com Inc. System and method for enabling users to interact in a virtual space
US6396509B1 (en) * 1998-02-21 2002-05-28 Koninklijke Philips Electronics N.V. Attention-based interaction in a virtual environment
WO2000070557A2 (en) * 1999-05-14 2000-11-23 Graphic Gems Method and apparatus for registering lots in a shared virtual world
US6772195B1 (en) * 1999-10-29 2004-08-03 Electronic Arts, Inc. Chat clusters for a virtual world application
US20070233785A1 (en) * 2006-03-30 2007-10-04 International Business Machines Corporation Communicating using collaboration spaces
EP1964597A1 (en) * 2007-03-01 2008-09-03 Sony Computer Entertainment Europe Ltd. Apparatus and method of modifying an online environment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013181026A1 (en) * 2012-06-02 2013-12-05 Social Communications Company Interfacing with a spatial virtual communications environment
CN107688418A (en) * 2017-05-05 2018-02-13 平安科技(深圳)有限公司 Network indicates the methods of exhibiting and system of control

Also Published As

Publication number Publication date
US20090288007A1 (en) 2009-11-19
WO2011016967A3 (en) 2011-04-14
KR20120050980A (en) 2012-05-21
CN102483819A (en) 2012-05-30
EP2460138A2 (en) 2012-06-06
IL217290A0 (en) 2012-02-29

Similar Documents

Publication Publication Date Title
US11785056B2 (en) Web browser interface for spatial communication environments
US20090288007A1 (en) Spatial interfaces for realtime networked communications
US9813463B2 (en) Phoning into virtual communication environments
USRE46309E1 (en) Application sharing
US9411489B2 (en) Interfacing with a spatial virtual communication environment
US8930472B2 (en) Promoting communicant interactions in a network communications environment
US8732593B2 (en) Shared virtual area communication environment based apparatus and methods
US20130174059A1 (en) Communicating between a virtual area and a physical space
US20230339816A1 (en) Visual Communications

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201080034697.4

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10806818

Country of ref document: EP

Kind code of ref document: A2

WWE Wipo information: entry into national phase

Ref document number: 9618/CHENP/2011

Country of ref document: IN

ENP Entry into the national phase

Ref document number: 20127002141

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2010806818

Country of ref document: EP