CA2858586A1 - Context-sensitive views - Google Patents
Context-sensitive views Download PDFInfo
- Publication number
- CA2858586A1 CA2858586A1 CA2858586A CA2858586A CA2858586A1 CA 2858586 A1 CA2858586 A1 CA 2858586A1 CA 2858586 A CA2858586 A CA 2858586A CA 2858586 A CA2858586 A CA 2858586A CA 2858586 A1 CA2858586 A1 CA 2858586A1
- Authority
- CA
- Canada
- Prior art keywords
- contextual
- data
- views
- modifying
- computer system
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
- G06F3/04855—Interaction with scrollbars
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/904—Browsing; Visualisation therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04812—Interaction techniques based on cursor appearance or behaviour, e.g. being affected by the presence of displayed objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04847—Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
Abstract
A context-sensitive viewing system is disclosed in which various data visualizations, also referred to a contextual views, of a common set of data may be viewed by a user on an electronic device. Data in the system may comprise data objects and associated properties and/or metadata, and may be stored in one or more electronic data stores. As a user of the system views and manipulates a first contextual view of a set of data objects, one or more other contextual views of the same set of data objects may be updated accordingly. Updates to the secondary contextual views may, in various embodiments, happen real-time. Further, the secondary contextual views may be visible to the user simultaneously with the primary contextual view. A
user may switch from one view to another, and may manipulate data in any view, resulting in updates in the other views.
user may switch from one view to another, and may manipulate data in any view, resulting in updates in the other views.
Description
CONTEXT-SENSITIVE VIEWS
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims a priority benefit under 35 U.S.C. 119 to U.S. Provisional Patent Application No. 61/864,048, filed on August 9, 2013, and titled "CONTEXT-SENSITIVE VIEWS," which is hereby incorporated by reference herein in its entirety.
TECHNICAL FIELD
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims a priority benefit under 35 U.S.C. 119 to U.S. Provisional Patent Application No. 61/864,048, filed on August 9, 2013, and titled "CONTEXT-SENSITIVE VIEWS," which is hereby incorporated by reference herein in its entirety.
TECHNICAL FIELD
[0002] The present disclosure relates to systems and techniques for data integration, analysis, and visualization. More specifically, the present disclosure relates to integration, analysis, and visualization of data objects in various contextual views.
BACKGROUND
BACKGROUND
[0003] Visualizations may enable faster and more thorough understandings of sets of data and information. Such visualizations of data and other information may be referred to as data visualizations. Data visualizations may, for example, visually transform and/or restructure data so as to provide new perspectives to a viewer of the visualization. A particular type of data visualization may be referred to as a contextual view. Examples of data visualizations include graphs, maps, tables, and/or lists, among others. Data visualizations may include displaying individual pieces of data in, for example, various arrangements, various sizes, various colors, and/or may include multi-dimensional aspects.
SUMMARY
SUMMARY
[0004] The systems, methods, and devices described herein each have several aspects, no single one of which is solely responsible for its desirable attributes. Without limiting the scope of this disclosure, several non-limiting features will now be discussed briefly.
[0005] A context-sensitive viewing system is disclosed in which various data visualizations, also referred to a contextual views, of a common set of data may be viewed by a user on an electronic device. Data in the context-sensitive viewing system may comprise data objects and associated properties and/or metadata. As a user of the system views and manipulates a first contextual view of a set of data objects, one or more other contextual views of the same set of data objects may be updated accordingly.
[0006] In various embodiments, a user of the context-sensitive viewing system may switch from a primary contextual view to a secondary contextual view, thereby making the switched-to contextual view the new primary contextual view.
Data objects may be manipulated in any view, resulting in updates in the other views. Context switching may be accomplished through inputs from the user. For example, the user may click on a preview of a secondary view, and/or may scroll from one view to the next.
Data objects may be manipulated in any view, resulting in updates in the other views. Context switching may be accomplished through inputs from the user. For example, the user may click on a preview of a secondary view, and/or may scroll from one view to the next.
[0007] The context-sensitive viewing system advantageously enables a user to view a particular set of data objects in multiple visualization contexts.
Previews of the set of data in other visualization may be quickly reviewed by the user to determine the most beneficial context for information extraction.
Further, manipulations by the user in one context are propagated to the other contexts, allowing fast analysis of the impacts of changes to the set of data.
Previews of the set of data in other visualization may be quickly reviewed by the user to determine the most beneficial context for information extraction.
Further, manipulations by the user in one context are propagated to the other contexts, allowing fast analysis of the impacts of changes to the set of data.
[0008] In an embodiment, a computer system is disclosed comprising one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least: an electronic database configured to store a plurality of data objects and properties associated with each of the data objects; and a context viewing module configured to:
generate a primary contextual view including a visualization of a set of data objects and associated properties; generate one or more secondary contextual views, each secondary contextual view including respective secondary visualizations of the set of data objects; receive a user input modifying the primary contextual view; and in response to receiving a user input modifying the primary contextual view, modify one or more of the secondary contextual views based at least in part on the user input.
generate a primary contextual view including a visualization of a set of data objects and associated properties; generate one or more secondary contextual views, each secondary contextual view including respective secondary visualizations of the set of data objects; receive a user input modifying the primary contextual view; and in response to receiving a user input modifying the primary contextual view, modify one or more of the secondary contextual views based at least in part on the user input.
[0009] According to an aspect, the context viewing module may be further configured to: in response to receiving a user input modifying the primary contextual view, determine which of the one or more of the secondary contextual views to modify.
[0010] According to an aspect, modifying one or more of the secondary contextual views may comprise modifying all of the one or more secondary contextual view.
[0011] According to an aspect, modifying one or more of the secondary contextual views may comprise modifying any of the one or more secondary contextual views that are currently viewable by a user.
[0012] According to an aspect, modifying one or more of the secondary contextual views may comprise modifying any of the one or more secondary contextual views that are immediately adjacent to the primary contextual view.
[0013] According to an aspect, each of the visualization and/or the secondary visualizations may include at least one of a graph, a map, a table, a timeline, a histogram, a list, a reader interface, or a postboard interface.
[0014] According to an aspect, the one or more secondary contextual views may comprise contextual previews.
[0015] According to an aspect, the one or more secondary contextual views may be viewable in a drawer or scrollbar on a user interface, and the one or more secondary contextual views may be selectable by a user.
[0016] According to an aspect, the one or more secondary contextual views may be substantially the same size as the primary contextual view.
[0017] According to an aspect, the one or more secondary contextual views may be configured to be accessible by a user through the use of a scrollbar.
[0018] According to an aspect, the scrollbar may include at least one of tick marks indicating the locations of the one or more secondary views or contextual previews accessible in pop up windows.
[0019] According to an aspect, the one or more secondary contextual views may be positioned laterally to the primary contextual view, and the secondary contextual views may be accessible by a user through a user input including at least one of a mouse cursor or a touch input.
[0020] According to an aspect, a user input modifying the primary contextual view may comprise at least one of adding data objects, removing data objects, modifying data objects, moving data objects, modifying properties associated with data objects, or modifying and/or manipulating links between data objects.
[0021] In an embodiment, a computer system is disclosed comprising one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least: a context viewing module configured to: display a first data visualization of a set of data objects and properties associated with data objects of the set of data objects;
provide one or more secondary data visualizations of the set of data objects;
receive a user input modifying the first data visualization; and in response to receiving a user input modifying the first data visualization, implement modifications to at least some of the one or more secondary data visualizations based at least in part on the user input.
provide one or more secondary data visualizations of the set of data objects;
receive a user input modifying the first data visualization; and in response to receiving a user input modifying the first data visualization, implement modifications to at least some of the one or more secondary data visualizations based at least in part on the user input.
[0022] According to an aspect, the context viewing module may be further configured to: in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are currently displayed to the user, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
[0023] According to an aspect, the context viewing module may be further configured to: in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are adjacent to the first data visualization, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
[0024] In an embodiment, a computer-implemented method of updating multiple contextual views is disclosed, the method comprising: providing an electronic database configured to store a plurality of data objects and metadata associated with each of the plurality of data objects; generating, by a computing system having one or more computer processors, based at least in part on the plurality of data objects and associated metadata, a primary contextual view and one or more secondary contextual views; receiving, via an input device of the computing system, a user input modifying the primary contextual view;
determining, by the computing system, based on the received user input, modifications of the one or more secondary contextual views that correspond to the modification of the primary contextual view; modifying at least some of the one or more secondary contextual views based on the determined modifications.
determining, by the computing system, based on the received user input, modifications of the one or more secondary contextual views that correspond to the modification of the primary contextual view; modifying at least some of the one or more secondary contextual views based on the determined modifications.
[0025] According to an aspect, the method may further comprise providing, on an electronic display of the computing system, the generated primary contextual view and one or more of the secondary contextual views in a preview drawer.
[0026] According to an aspect, the method may further comprise providing, on an electronic display of the computing system, the generated primary contextual view; and providing, on the electronic display of the computer system, a scrollbar that enables a user to scroll to any of the one or more of the secondary contextual views and view any of the one or more of the secondary contextual views on the electronic display.
[0027] According to an aspect, modifying at least some of the one or more secondary contextual views based on the determined modifications may comprise modifying any secondary contextual views that are immediately viewable by a user.
BRIEF DESCRIPTION OF THE DRAWINGS
BRIEF DESCRIPTION OF THE DRAWINGS
[0028] FIG. 1A illustrates a user interface of a context-sensitive viewing system, including relationships described in a data store using a dynamic ontology, according to an embodiment of the present disclosure.
[0029] FIG. 1B
illustrates a user interface of the context-sensitive viewing system including an expanded drawer with various contextual views, according to an embodiment of the present disclosure.
illustrates a user interface of the context-sensitive viewing system including an expanded drawer with various contextual views, according to an embodiment of the present disclosure.
[0030] FIGS. 10 and 1D illustrate user interfaces of the context-sensitive viewing system in which selected data objects are updated in various contextual views, according to embodiments of the present disclosure.
[0031] FIGS. 2A and 2B illustrate another user interface of the context-sensitive viewing system including multiple contextual views, according to embodiments of the present disclosure.
[0032] FIGS. 20 and 2D illustrate scrollbar aspects of user interfaces of the context-sensitive viewing system, according to embodiments of the present disclosure.
[0033] FIG. 3 is a flowchart depicting an illustrative operation of the context-sensitive viewing system, according to an embodiment of the present disclosure.
[0034] FIG. 4A
illustrates a network environment in which the context-sensitive viewing system may operate, according to an embodiment of the present disclosure.
illustrates a network environment in which the context-sensitive viewing system may operate, according to an embodiment of the present disclosure.
[0035] FIG. 4B
illustrates a computer system with which certain methods discussed herein may be implemented, according to embodiments of the present disclosure.
illustrates a computer system with which certain methods discussed herein may be implemented, according to embodiments of the present disclosure.
[0036] FIG. 5 illustrates an embodiment of a database system using an ontology.
[0037] FIG. 6 illustrates an embodiment of a system for creating data in a data store using a dynamic ontology.
DETAILED DESCRIPTION OF SPECIFIC EMBODIMENTS
Overview [0038] A context-sensitive viewing system is disclosed in which various data visualizations, also referred to a contextual views, of a common set of data may be viewed by a user on an electronic device. Data in the context-sensitive viewing system may comprise data objects and associated properties and/or metadata, and may be stored in one or more electronic data stores. As a user of the system views and manipulates a first contextual view (also referred to as the "primary contextual view") of a set of data objects, one or more other contextual views (also referred to as "secondary contextual views") of the same set of data objects may be updated accordingly.
DETAILED DESCRIPTION OF SPECIFIC EMBODIMENTS
Overview [0038] A context-sensitive viewing system is disclosed in which various data visualizations, also referred to a contextual views, of a common set of data may be viewed by a user on an electronic device. Data in the context-sensitive viewing system may comprise data objects and associated properties and/or metadata, and may be stored in one or more electronic data stores. As a user of the system views and manipulates a first contextual view (also referred to as the "primary contextual view") of a set of data objects, one or more other contextual views (also referred to as "secondary contextual views") of the same set of data objects may be updated accordingly.
[0039]
Updates to the secondary contextual views may, in various embodiments, happen real-time or may happen upon the occurrence of a triggering event (for example, a user input). In various embodiments, the secondary contextual views may comprise previews and/or thumbnails. Further, the secondary contextual views may be visible to the user simultaneously with the primary contextual view.
For example, the user of the context-sensitive viewing system may view a particular set of data objects in multiple visualization contexts. Further, as the user updates the set of data objects in one context, the set of data objects may automatically be updated in one or more secondary contexts.
Updates to the secondary contextual views may, in various embodiments, happen real-time or may happen upon the occurrence of a triggering event (for example, a user input). In various embodiments, the secondary contextual views may comprise previews and/or thumbnails. Further, the secondary contextual views may be visible to the user simultaneously with the primary contextual view.
For example, the user of the context-sensitive viewing system may view a particular set of data objects in multiple visualization contexts. Further, as the user updates the set of data objects in one context, the set of data objects may automatically be updated in one or more secondary contexts.
[0040]
For the sake of brevity, contextual views may be referred to herein simply as "views" or "contexts." For example, a primary contextual view may be referred to as a "primary view." Additionally, the terms "contextual view" and "data visualization" may be used interchangeably.
For the sake of brevity, contextual views may be referred to herein simply as "views" or "contexts." For example, a primary contextual view may be referred to as a "primary view." Additionally, the terms "contextual view" and "data visualization" may be used interchangeably.
[0041] In various ways and in various embodiments, a user of the context-sensitive viewing system may switch from a primary contextual view to a secondary contextual view, thereby making the switched-to contextual view the new primary contextual view. Data objects may be manipulated in any view, resulting in updates in the other views. Context switching may be accomplished through inputs from the user. For example, the user may click on a preview of a secondary view, and/or may scroll from one view to the next.
[0042]
Examples of contextual views (and/or data visualizations) of the context-sensitive viewing system include, but are not limited to graphs, maps, tables, timelines, histograms, and/or lists, among other types of data visualizations.
In an embodiment, a contextual view comprises a graph of connected data objects as described below. In an embodiment, a contextual view comprises an interactive mapping application, an example of which is described in U.S. Patent Appl. No.
13/917,571 filed on June 13, 2013, and titled "Interactive Geospatial Map,"
which is hereby incorporated by reference herein in its entirety and for all purposes.
In an embodiment, a contextual view comprises a reader interface that enables a user to review large amounts of notes and other textual information. An example of such a reader interface is described in U.S. Provisional Patent Application No.
61/863,792, filed on August 8, 2013, and titled "Cable Reader Labeling,", which is hereby incorporated by reference herein in its entirety and for all purposes. In an embodiment, a contextual view comprises a postboard view in which notes and textual clips may be listed, an example of which is described in U.S.
Provisional Patent Application No. 61/863,814, filed on August 8, 2013, and titled "Cable Reader Snippets and Postboard," which is hereby incorporated by reference herein in its entirety and for all purposes. In an embodiment, a contextual view comprises a time series graph, timeline, and/or histogram, examples of which are described in U.S. Patent No. 8,280,880, titled "Generating Dynamic Date Sets That Represent Market Conditions," and U.S. Patent No. 8,280,880, titled "Filter Chains With Associated Views For Exploring Large Data Sets," each of which is hereby incorporated by reference herein in its entirety and for all purposes.
Definitions [0043] In order to facilitate an understanding of the systems and methods discussed herein, a number of terms are defined below. The terms defined below, as well as other terms used herein, should be construed to include the provided definitions, the ordinary and customary meaning of the terms, and/or any other implied meaning for the respective terms. The definitions below do not limit the meaning of these terms, but only provide exemplary definitions.
Examples of contextual views (and/or data visualizations) of the context-sensitive viewing system include, but are not limited to graphs, maps, tables, timelines, histograms, and/or lists, among other types of data visualizations.
In an embodiment, a contextual view comprises a graph of connected data objects as described below. In an embodiment, a contextual view comprises an interactive mapping application, an example of which is described in U.S. Patent Appl. No.
13/917,571 filed on June 13, 2013, and titled "Interactive Geospatial Map,"
which is hereby incorporated by reference herein in its entirety and for all purposes.
In an embodiment, a contextual view comprises a reader interface that enables a user to review large amounts of notes and other textual information. An example of such a reader interface is described in U.S. Provisional Patent Application No.
61/863,792, filed on August 8, 2013, and titled "Cable Reader Labeling,", which is hereby incorporated by reference herein in its entirety and for all purposes. In an embodiment, a contextual view comprises a postboard view in which notes and textual clips may be listed, an example of which is described in U.S.
Provisional Patent Application No. 61/863,814, filed on August 8, 2013, and titled "Cable Reader Snippets and Postboard," which is hereby incorporated by reference herein in its entirety and for all purposes. In an embodiment, a contextual view comprises a time series graph, timeline, and/or histogram, examples of which are described in U.S. Patent No. 8,280,880, titled "Generating Dynamic Date Sets That Represent Market Conditions," and U.S. Patent No. 8,280,880, titled "Filter Chains With Associated Views For Exploring Large Data Sets," each of which is hereby incorporated by reference herein in its entirety and for all purposes.
Definitions [0043] In order to facilitate an understanding of the systems and methods discussed herein, a number of terms are defined below. The terms defined below, as well as other terms used herein, should be construed to include the provided definitions, the ordinary and customary meaning of the terms, and/or any other implied meaning for the respective terms. The definitions below do not limit the meaning of these terms, but only provide exemplary definitions.
[0044] Ontology: Stored information that provides a data model for storage of data in one or more databases. For example, the stored data may comprise definitions for object types and property types for data in a database, and how objects and properties may be related.
[0045] Database: A broad term for any data structure for storing and/or organizing data, including, but not limited to, relational databases (Oracle database, mySQL database, etc.), spreadsheets, XML files, and text file, among others.
[0046] Data Object or Object: A data container for information representing specific things in the world that have a number of definable properties.
For example, a data object can represent an entity such as a person, a place, an organization, a market instrument, or other noun. A data object can represent an event that happens at a point in time or for a duration. A data object can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object may be associated with a unique identifier that uniquely identifies the data object. The object's attributes (e.g.
metadata about the object) may be represented in one or more properties.
For example, a data object can represent an entity such as a person, a place, an organization, a market instrument, or other noun. A data object can represent an event that happens at a point in time or for a duration. A data object can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object may be associated with a unique identifier that uniquely identifies the data object. The object's attributes (e.g.
metadata about the object) may be represented in one or more properties.
[0047] Object Type: Type of a data object (e.g., Person, Event, or Document). Object types may be defined by an ontology and may be modified or updated to include additional object types. An object definition (e.g., in an ontology) may include how the object is related to other objects, such as being a sub-object type of another object type (e.g. an agent may be a sub-object type of a person object type), and the properties the object type may have.
[0048] Properties: Attributes of a data object that represent individual data items. At a minimum, each property of a data object has a property type and a value or values.
[0049] Property Type: The type of data a property is, such as a string, an integer, or a double. Property types may include complex property types, such as series data values associated with timed ticks (e.g. a time series), etc.
[0050] Property Value: The value associated with a property, which is of the type indicated in the property type associated with the property. A
property may have multiple values.
property may have multiple values.
[0051] Link: A connection between two data objects, based on, for example, a relationship, an event, and/or matching properties. Links may be directional, such as one representing a payment from person A to B, or bidirectional.
[0052] Link Set: Set of multiple links that are shared between two or more data objects.
[0053] Contextual view, context, view, data representation: A visual representation of data that may include various organizations, transformations, and/or restructuring of data so as to provide new perspectives to a viewer of the visualization. Examples of contexts include graphs, maps, tables, timelines, histograms, and/or lists, among others. Contextual views may include displaying individual pieces of data in, for example, various arrangements, various sizes, various colors, and/or may include multi-dimensional aspects., Contextual views may enable faster and more thorough understandings of sets of data and information.
Example User Interfaces [0054] FIG. 1A illustrates a user interface 101 of a context-sensitive viewing system, according to an embodiment of the present disclosure. The user interface 101 may be displayed on, for example, an electronic display of the system, such as client device 402 of Figure 4. Further, as is described in detail below, a user may provide inputs to the system through, for example, a mouse and pointer or through a touch interface.
Example User Interfaces [0054] FIG. 1A illustrates a user interface 101 of a context-sensitive viewing system, according to an embodiment of the present disclosure. The user interface 101 may be displayed on, for example, an electronic display of the system, such as client device 402 of Figure 4. Further, as is described in detail below, a user may provide inputs to the system through, for example, a mouse and pointer or through a touch interface.
[0055] The example user interface 101 includes a particular graphical contextual view and/or data visualization 103 of various data objects and relationships between those data objects. In the example user interface 101 of FIG. 1A, a single contextual view is provided that shows, properties, data objects, and links/relationships between the data objects using a graphical user interface (GUI). The data objects shown on the user interface 101 include, for example, person objects 121, 122, 123, 124, 125, and 126; a flight object 127; a financial account 128; and a computer object 129. The data objects are represented by nodes. The relationships and/or links between the various data objects of user interface 101 are represented by lines, and include links 104, 105, 106, 107, 108, 109, 110, 111, 112, and 113.
[0056] In this example contextual view 103, each person node (associated with person data objects), flight node (associated with flight data objects), financial account node (associated with financial account data objects), and computer node (associated with computer data objects) may have relationships and/or links with any of the other nodes through, for example, other objects such as payment objects.
As is described in detail in reference to FIGS. 4A and 4B below, the various data objects, data object properties, and/or relationships among those data objects and properties may be stored in, and retrieved from, one or more data stores and/or databases. As is further described in detail in reference to FIGS. 5 and 6 below, the data objects data object properties, and/or relationships may be stored using a dynamic ontology.
As is described in detail in reference to FIGS. 4A and 4B below, the various data objects, data object properties, and/or relationships among those data objects and properties may be stored in, and retrieved from, one or more data stores and/or databases. As is further described in detail in reference to FIGS. 5 and 6 below, the data objects data object properties, and/or relationships may be stored using a dynamic ontology.
[0057] Turning back to FIG. 1A, various example relationships between data objects are represented. For example, relationship 104 is based on a payment associated with the individuals indicated in person data objects 121 and 123.
The link 104 represents these shared payments (for example, the individual associated with data object 121 may have paid the individual associated with data object on three occasions). The relationship is further indicated by the common relationship between person data objects 121 and 123 and financial account data object 128. For example, link 111 indicates that person data object 121 transferred money into financial account data object 128, while person data object 123 transferred money out of financial account data object 128. In another example, the relationships between person data objects 124 and 125 and flight data object are indicated by links 106, 109, and 110. In this example, person data objects and 125 have a common address and were passengers on the same flight data object 127. In an embodiment, further details related to the relationships between the various objects may be displayed. For example, links 111 and 112 may, in some embodiments, indicate the timing of the respective money transfers. In another example, the time of the flight associated with the flight data object 127 may be shown.
The link 104 represents these shared payments (for example, the individual associated with data object 121 may have paid the individual associated with data object on three occasions). The relationship is further indicated by the common relationship between person data objects 121 and 123 and financial account data object 128. For example, link 111 indicates that person data object 121 transferred money into financial account data object 128, while person data object 123 transferred money out of financial account data object 128. In another example, the relationships between person data objects 124 and 125 and flight data object are indicated by links 106, 109, and 110. In this example, person data objects and 125 have a common address and were passengers on the same flight data object 127. In an embodiment, further details related to the relationships between the various objects may be displayed. For example, links 111 and 112 may, in some embodiments, indicate the timing of the respective money transfers. In another example, the time of the flight associated with the flight data object 127 may be shown.
[0058] Relationships between data objects may be stored as links, or in some embodiments, as properties, where a relationship may be detected between the properties. In some cases the links may be directional. For example, a payment link may have a direction associated with the payment, where one person object is a receiver of a payment, and another person object is the payer of payment.
[0059] In addition to visually showing graphical data visualization 103, the user interface 101 may allow various manipulations. For example, the various data objects of the context-sensitive viewing system may be searched using a search interface 102 (e.g., text string matching of object properties), inspected (e.g., properties and associated data viewed), filtered (e.g., narrowing the universe of objects into sets and subsets by properties or relationships), and statistically aggregated (e.g., numerically summarized based on summarization criteria), among other operations and visualizations. Further, the various data objects represented in the data visualization 103 may be moved, accessed, deleted from the interface, among other manipulations. Additional data objects and associated links may be added to the data visualization 103, and exiting data objects and links may be edited and/or otherwise altered.
[0060] The user interface 101 further includes a user-accessible drawer 130. The drawer 130 may be opened or closed by a user of the context-sensitive viewing system. The drawer 130 is shown in a closed state in FIG.
1A. As indicated by an upward pointing arrow, the drawer 130 may be opened by a user clicking and dragging the drawer open and/or touching and pulling the drawer open, as show in FIG. 1B.
1A. As indicated by an upward pointing arrow, the drawer 130 may be opened by a user clicking and dragging the drawer open and/or touching and pulling the drawer open, as show in FIG. 1B.
[0061] FIG. 1B illustrates the user interface 101 of the context-sensitive viewing system including an expanded drawer with various contextual views, according to an embodiment of the present disclosure. While including many of the same objects as FIG. 1A, the graphical data visualization 103 of FIG. 1B is simplified for the sake of clarity. As shown, the user has used the mouse pointer 132 to pull open the drawer 130, revealing the contents of the drawer. The drawer includes indications of various available contextual views 152, 154, 156, 158, and 160. Arrows 134 and 136 indicate that, in some embodiment, additional indications of available views may extend beyond those that are currently visible. A
user of the context-sensitive viewing system may, in an embodiment, scroll through the indications of views.
user of the context-sensitive viewing system may, in an embodiment, scroll through the indications of views.
[0062] In the user interface 101 of FIG. 1B, view indicators include reader view 154, map view 156, and postboard view 158. A user may click on or otherwise select one of the indicators 154, 156, or 158. Selecting one of the views has the effect of changing from the current primary graphical data visualization 103 to the selected view. When the user changes to a different view, the same underlying data set (including data objects, properties, links, and the line) will be used in generating the new view. For example, the user may select the "map" indicator 156, at which point graphical data visualization 103 would be replaced with a map contextual view in which the same data objects (121, 122, 123, 124, 125, 126) may, for example, be placed on the map based on location data associated with the respective data objects.
[0063] In an embodiment, the user may provide an input that causes the indicators 152, 154, 156, 158, 160 to be replaced with previews of the respective contexts (as shown in FIG. 1C). In another embodiment, when the user opens the drawer 130, previews of the respective contexts may be provided automatically.
[0064] FIGS. 1C illustrates the user interface 101 of the context-sensitive viewing system in which selected data objects are updated in various contextual views, according to an embodiment of the present disclosure. As shown in FIG.
1C, previews and/or thumbnails 154, 156, 158 are provided giving the user a preview of the current data in a reader, map, and postboard context, respectively. The context previews 154, 156, 158 include actual information derived from the data objects currently viewed in the graphical contextual view 103. For example, the map context preview 156 includes the locations of the various person data objects plotted on the map (see, for example, 123 person location 167). The reader context preview includes cables and/or notes related to the data objects and/or links currently shown in the graphical contextual view 103 (see, for example, payment tracked information 166). The postboard context preview 158 includes clippings and/or other user-generated notes associated with the data objects. In various embodiments other contextual previews may be displayed to the user.
1C, previews and/or thumbnails 154, 156, 158 are provided giving the user a preview of the current data in a reader, map, and postboard context, respectively. The context previews 154, 156, 158 include actual information derived from the data objects currently viewed in the graphical contextual view 103. For example, the map context preview 156 includes the locations of the various person data objects plotted on the map (see, for example, 123 person location 167). The reader context preview includes cables and/or notes related to the data objects and/or links currently shown in the graphical contextual view 103 (see, for example, payment tracked information 166). The postboard context preview 158 includes clippings and/or other user-generated notes associated with the data objects. In various embodiments other contextual previews may be displayed to the user.
[0065] As described below in reference to FIG. 3, in various embodiments, the contextual previews may be generated and/or updated when the user opens the drawer 130, before the user opens the drawer 130, when particular previews become visible to the user, when any (or certain) changes are made to one of the views, or at various other times. In some embodiments, one or more contextual views or previews may be updated based on particular criteria, for example, whether the view/preview is visible to the user, whether the view/preview is adjacent to one visible by the user, and the like.
[0066] FIG. 1C additionally shows that, in some embodiments, a preview may be updated when the user drags selected objects onto the preview itself.
In the example of FIG. 1C, the user has made selection 164, including data objects and 122. Using mouse cursor 162, the user drags the selected objects to the map context preview 156, at which point the map context preview is updated to include the selected objects.
In the example of FIG. 1C, the user has made selection 164, including data objects and 122. Using mouse cursor 162, the user drags the selected objects to the map context preview 156, at which point the map context preview is updated to include the selected objects.
[0067] FIGS. 1D illustrates the user interface 101 of the context-sensitive viewing system in which added data objects are updated in various contextual views, according to an embodiment of the present disclosure. In the embodiment of FIG. 10, the user has added person data object 114 to the graphical context/data visualization 103. The added data object 114 includes a link 115 to person data object 121, indicating a payment took place between the objects. By adding the data object 114, the three contextual previews 154, 156, and 158 are automatically updated to reflect the new data object. For example, a location 172 associated with data object 114 is added to the map preview 156, while information 170 regarding the payment link 115 is added to the reader preview 154.
[0068] In various embodiments other types of contextual previews may be provided. For example, a timeline context may be provided in which the various event associated with the data objects of graphical primary contextual view 103 may be mapped. In various embodiments, any other types of changes to the primary contextual view may be reflected in the secondary contextual views/previews.
For example, removing data object, editing data objects or properties, and the like.
For example, removing data object, editing data objects or properties, and the like.
[0069] In various embodiments, the drawer 130 may appear at different locations on the user interface 101, and/or may be a different size. For example, the drawer 130 may appear at a top of the user interface, or on either side of the user interface. In an embodiment, the location, size, and/or other appearance of the drawer 130 may be user-configurable.
[0070] FIGS. 2A and 2B illustrate another example user interface of the context-sensitive viewing system including multiple contextual views, according to embodiments of the present disclosure. FIG. 2A includes a display 202 (for example, an electronic display associated with an electronic device), a scrollbar 205, and contextual views 204, 206, 208, and 210 corresponding to a graphical view, reader view, map view, and postboard view, respectively. A user of the context-sensitive viewing system views the primary graphical view 204 on the display 202, while the other views are not visible. The user may scroll and/or swipe (as with a touch interface) between the various views, as indicated by arrows 212, 214.
The user may optionally use scrollbar 205 for scrolling between views.
The user may optionally use scrollbar 205 for scrolling between views.
[0071] FIG. 2B illustrates the user using mouse cursor 216 to scroll from the graphical view 204 to the reader view 206. Additional or fewer views may be available to the user. Additionally, in various embodiments multiple views may be arranged in other dimensions. For example, views may be laid out in a two-dimensional arrangement of rows and columns, and the display 202 may include a vertical scrollbar in addition to the horizontal scrollbar. In another embodiment, multiple views may be arranged in a three-dimensional arrangement, or any other arrangement suitable for user interaction with the context-sensitive viewing system.
[0072] In the embodiment of FIGS. 2A-2B, the user may interact with and/or manipulate the current, primary view as described above. Changes to the primary view are reflected in the secondary, non-visible views automatically, as described above with respect to the contextual previews of FIGS. 2C-2D.
Additionally, as described above, in various embodiments changes may not be implemented in the non-visible views until they become visible to the user.
Alternatively, adjacent views to the primary view may be updated more frequently (e.g. in realtime as changes are made to the primary view), while changes to non-adjacent views are made less frequently or not at all (e.g., a view may not be updated until it is adjacent to the primary view). In other embodiments, all views may be updated concurrently, such as in realtime, periodically, or in response to certain events.
Additionally, as described above, in various embodiments changes may not be implemented in the non-visible views until they become visible to the user.
Alternatively, adjacent views to the primary view may be updated more frequently (e.g. in realtime as changes are made to the primary view), while changes to non-adjacent views are made less frequently or not at all (e.g., a view may not be updated until it is adjacent to the primary view). In other embodiments, all views may be updated concurrently, such as in realtime, periodically, or in response to certain events.
[0073] FIGS. 2C and 2D illustrate alternative scrollbars that may be implemented in the context-sensitive viewing system, according to embodiments of the present disclosure. The scrollbars of FIGS. 20 and 2D may be implemented, for example, in the place of scrollbar 205 of FIGS. 2A and 2B.
[0074] In FIG. 2C the scrollbar 220 includes markers (or tick marks) 222 that indicate the locations of the various contextual views. Additionally, when mouse cursor 218 is placed over the scrollbar 220, a contextual preview 224 pops up.
The preview 224 thereby provides the user with an indication of the particular view available at that location. In various embodiments, either markers 222 or pop up contextual preview 224, or both or neither, may be included in scrollbar 220.
The preview 224 thereby provides the user with an indication of the particular view available at that location. In various embodiments, either markers 222 or pop up contextual preview 224, or both or neither, may be included in scrollbar 220.
[0075] In FIG. 2D the scrollbar includes contextual previews 230, 232, 234, and 236. A box indicator 228 indicates the currently viewable contextual view.
For example, in FIG. 2D a graphical view corresponding to preview 232 is visible to the user. The user may use the mouse cursor 226 to move the indicator 228 along the scrollbar.
For example, in FIG. 2D a graphical view corresponding to preview 232 is visible to the user. The user may use the mouse cursor 226 to move the indicator 228 along the scrollbar.
[0076] In an embodiment, multiple of the same type of contextual view may be accessible to the user. For example, timeline contextual views may be available.
Example Operations [0077] FIG. 3 is a flowchart depicting an illustrative operation of the context-sensitive viewing system, according to an embodiment of the present disclosure. In various embodiments, fewer blocks or additional blocks may be included in the process, or various blocks may be performed in an order different from that shown in FIG. 3. The various operations and functions described in FIG. 3 may, in various embodiments, be implemented by one or more modules of the context-sensitive viewing system.
Example Operations [0077] FIG. 3 is a flowchart depicting an illustrative operation of the context-sensitive viewing system, according to an embodiment of the present disclosure. In various embodiments, fewer blocks or additional blocks may be included in the process, or various blocks may be performed in an order different from that shown in FIG. 3. The various operations and functions described in FIG. 3 may, in various embodiments, be implemented by one or more modules of the context-sensitive viewing system.
[0078] At block 302, the context-sensitive viewing system receives user input at the first contextual view and/or data visualization user interface.
In this present example, the first contextual view comprises the primary view with which the user is currently interacting. User inputs may include, for example, adding and/or deleting data objects, manipulating data objects, altering and/or editing data object properties and/or links, among other inputs.
In this present example, the first contextual view comprises the primary view with which the user is currently interacting. User inputs may include, for example, adding and/or deleting data objects, manipulating data objects, altering and/or editing data object properties and/or links, among other inputs.
[0079] At block 304, updates to the contextual view are determined by the context-sensitive viewing system based on the user input. For example, if the user provides an input to add a data object to the view, information associated with the data object to be added may be retrieved from a particular data store. The retrieved data object may then be displayed to the user and/or otherwise represented on the first/primary contextual view.
[0080] At block 306, the user may optionally select a different contextual view. Selecting a second contextual view may be accomplished in any of the ways described above in reference to the user interfaces of FIGS. 1A-1D and 2A-2D.
For example, the user may drag objects from the first contextual view to a second contextual view. Alternatively, the user may click on or select a second contextual view. In another example, the user may scroll and/or slide the user interface to a second contextual view. In yet another embodiment, the user may drag one or more data objects onto another contextual view (as in the example of FIG. 1C). In an embodiment, block 306 may be optional. For example, in an embodiment, block may follow block 304 without any user action.
For example, the user may drag objects from the first contextual view to a second contextual view. Alternatively, the user may click on or select a second contextual view. In another example, the user may scroll and/or slide the user interface to a second contextual view. In yet another embodiment, the user may drag one or more data objects onto another contextual view (as in the example of FIG. 1C). In an embodiment, block 306 may be optional. For example, in an embodiment, block may follow block 304 without any user action.
[0081] At block 308, similar to block 304, updates to other contextual views may optionally be determined and displayed to the user. For example, when the user adds a data object to the first/primary contextual view, the same data object may be added to one or more other contextual views of the context-sensitive viewing system, as appropriate. In an example, when the user adds a person data object to a first graphical contextual view, the same person data object may be added to one or more other graphical contextual views. Further, the location(s) associated with that person data object may be added to one or more other map-based contextual views. Additionally, cables or other information, and/or user-generated snippets or notes associated with that person data object may be added to one or more other relevant contextual views.
[0082] The particular other contextual views that may be updated may depend on, for example, the particular implementation of the context-sensitive viewing system, user settings, and/or processing capability of the system. In an embodiment, at block 310, all other contextual views are updated simultaneously with, or very soon after, the updating of the first contextual. In another embodiment, at block 312, contextual views that are adjacent to the first view may be updated when the first view is updated. For example, in the embodiment of FIG. 2A, views that are immediately adjacent to the first, currently active view may be updated.
Alternatively, views that are near the current view, for example, two or three away from the current view, may be updated. In yet another embodiment, at block 314, any visible contextual views may be updated. For example, in the preview drawer embodiment of FIG. 1D, or the scrollbar of FIG. 2D, any previews that are currently visible on the display may be updated. Alternatively, visible previews may be updated when the drawer is opened. In another alternative, hidden previews when the drawer is closed may be updated such that the updated previews may be visible when the drawer is opened. In another example, as in the embodiment of FIG.
2B, as the user scrolls or slides from one view to the next, the next view may be updated when it becomes visible to the user. In another example, as in the embodiment of FIG. 2C, the popup preview may be updated when it becomes visible to the user.
Any combination of the above described examples may be implemented in the context-sensitive viewing system.
Alternatively, views that are near the current view, for example, two or three away from the current view, may be updated. In yet another embodiment, at block 314, any visible contextual views may be updated. For example, in the preview drawer embodiment of FIG. 1D, or the scrollbar of FIG. 2D, any previews that are currently visible on the display may be updated. Alternatively, visible previews may be updated when the drawer is opened. In another alternative, hidden previews when the drawer is closed may be updated such that the updated previews may be visible when the drawer is opened. In another example, as in the embodiment of FIG.
2B, as the user scrolls or slides from one view to the next, the next view may be updated when it becomes visible to the user. In another example, as in the embodiment of FIG. 2C, the popup preview may be updated when it becomes visible to the user.
Any combination of the above described examples may be implemented in the context-sensitive viewing system.
[0083] In an embodiment, updating of other contextual views is determined based on processing capability available to the context-sensitive viewing system. For example, additional contextual views may be updated when more processing capability is available. In another example, updates to particular contextual views may be delayed until visible to the user so as to reduce power consumption and/or processing power.
Implementation Mechanisms [0084] Turning to FIG. 4A, an illustrative network environment 400 in which the context-sensitive viewing system may operate, according to an embodiment of the present disclosure, is shown. The network environment 400 may include a client device 402, a network 408, a server device 404, and a database 406. The constituents of the network environment 400 may be in communication with each other either locally or over the network 408.
Implementation Mechanisms [0084] Turning to FIG. 4A, an illustrative network environment 400 in which the context-sensitive viewing system may operate, according to an embodiment of the present disclosure, is shown. The network environment 400 may include a client device 402, a network 408, a server device 404, and a database 406. The constituents of the network environment 400 may be in communication with each other either locally or over the network 408.
[0085] The client device 402 may be any computing device capable of receiving input and providing output to a user. For example, the client device may provide a contextual view of a data visualization to the user, among other functions. The client device 402 may also be capable of communicating over the network 408, for example, to request data objects, data visualization information, and/or contextual view information from the server device 404. In some embodiments, the client device 402 may include non-transitory computer-readable medium storage for storing data objects, data visualization information, and/or contextual view information. In an embodiment, the context-sensitive viewing system may include a plurality of client devices, each of which may communicate with each other, and with the network 408.
[0086] The network 408 may be any wired network, wireless network, or combination thereof. In addition, the network 408 may be a personal area network, local area network, wide area network, cable network, satellite network, cellular telephone network, or combination thereof. Protocols and components for communicating via the Internet or any of the other aforementioned types of communication networks are well known to those skilled in the art of computer communications and thus, need not be described in more detail herein.
[0087] The server device 404 is a computing device that may perform a variety of tasks to implement the contextual views and data visualizations of the context-sensitive viewing system. For example, the server device 404 may generate a user interface, including various contextual views, for display to the user via the client device 402. Alternatively, the server device 404 may receive requests for data and/or data objects from the client device 402, and may provide the requested data to the client device 402. The server device 404 may also generate requested data visualizations and/or contextual views that may be transmitted over the network 408, and provided to the user via the client device 402. Additional operations of the server device 404 and/or the client device 402 are described in further detail with respect to FIG. 4B.
[0088] The server device 404 may be in communication with the database 406. The database 406 may store one or more data objects, data visualization information, and/or contextual view information. The database 406 may be embodied in hard disk drives, solid state memories, and/or any other type of non-transitory, computer-readable storage medium remotely or locally accessible to the server device 404. The database 406 may also be distributed or partitioned across multiple storage devices as is known in the art without departing from the spirit and scope of the present disclosure.
[0089] FIG. 4A further illustrates that client device 402 includes a client-side context viewing module 410 and server device 404 includes a server-side context viewing module 412. Either of the context viewing modules 410, 412 may be configured to implement the various aspects of the context-sensitive viewing system as described herein. For example, in various embodiments, either of the context viewing modules 410, 412 may implement the user interface and/or various data visualizations and contextual views of the context-sensitive viewing system.
As described in further detail in reference to FIG. 4B below, context viewing modules 410, 412 may comprise software, and may be stored in one or more computer-readable media of the client device 402 and/or server device 404. In other embodiments, context viewing modules 410, 412 may comprise hardware modules.
Further, in various embodiments, context viewing modules 410, 412 may comprise additional modules that may implement the functionality of the context-sensitive viewing system. Functionality discussed above with reference to the client device and the server device may be changed to other devices (e.g., functions discussed with reference to the server device may be performed by the client device and/or vice versa) in various embodiments.
As described in further detail in reference to FIG. 4B below, context viewing modules 410, 412 may comprise software, and may be stored in one or more computer-readable media of the client device 402 and/or server device 404. In other embodiments, context viewing modules 410, 412 may comprise hardware modules.
Further, in various embodiments, context viewing modules 410, 412 may comprise additional modules that may implement the functionality of the context-sensitive viewing system. Functionality discussed above with reference to the client device and the server device may be changed to other devices (e.g., functions discussed with reference to the server device may be performed by the client device and/or vice versa) in various embodiments.
[0090] According to various embodiments, the techniques described herein may be implemented by one or more special-purpose computing devices.
The special-purpose computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the techniques, or may include one or more general purpose hardware processors programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the techniques. The special-purpose computing devices may be desktop computer systems, server computer systems, portable computer systems, handheld devices, networking devices or any other device or combination of devices that incorporate hard-wired and/or program logic to implement the techniques.
The special-purpose computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as one or more application-specific integrated circuits (ASICs) or field programmable gate arrays (FPGAs) that are persistently programmed to perform the techniques, or may include one or more general purpose hardware processors programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such special-purpose computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the techniques. The special-purpose computing devices may be desktop computer systems, server computer systems, portable computer systems, handheld devices, networking devices or any other device or combination of devices that incorporate hard-wired and/or program logic to implement the techniques.
[0091] Computing device(s), such as the client device 402 and/or the server device 404, are generally controlled and coordinated by operating system software, such as i0S, Android, Chrome OS, Windows XP, Windows Vista, Windows 7, Windows 8, Windows Server, Windows CE, Unix, Linux, SunOS, Solaris, i0S, Blackberry OS, VxWorks, or other compatible operating systems.
In other embodiments, a computing device may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface ("GUI"), among other things.
In other embodiments, a computing device may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface functionality, such as a graphical user interface ("GUI"), among other things.
[0092] For example, FIG. 4B is a block diagram that illustrates a computer system 420 upon which systems and methods discussed herein may be implemented. Computer system 420 includes a bus 422 or other communication mechanism for communicating information, and a hardware processor, or multiple processors, 424 coupled with bus 422 for processing information. Hardware processor(s) 424 may be, for example, one or more general purpose microprocessors. In various embodiments, one or more components of the computer system 420 may be included in the client device 402 and/or the server device 404 of FIG. 4A. In an embodiment, the client device 402 may be embodied in the computer system 420, while in another embodiment the server device 404 may be embodied in the computer system 420.
[0093] Computer system 420 also includes a main memory 426, such as a random access memory (RAM), cache and/or other dynamic storage devices, coupled to bus 422 for storing information and instructions to be executed by processor 424. Main memory 426 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 424. Such instructions, when stored in storage media accessible to processor 424, render computer system 420 into a special-purpose machine that is customized to perform the operations specified in the instructions.
[0094] Computer system 420 further includes a read only memory (ROM) 428 or other static storage device coupled to bus 422 for storing static information and instructions for processor 424. A storage device 430, such as a magnetic disk, optical disk, or USB thumb drive (Flash drive), etc., is provided and coupled to bus 422 for storing information and instructions.
[0095] Computer system 420 may be coupled via bus 422 to a display 432, such as a cathode ray tube (CRT) or LCD display (or touch screen), for displaying information to a computer user. An input device 434, including alphanumeric and other keys, is coupled to bus 422 for communicating information and command selections to processor 424. Another type of user input device is cursor control 436, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 424 and for controlling cursor movement on display 432. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. In some embodiments, the same direction information and command selections as cursor control may be implemented via receiving touches on a touch screen without a cursor.
[0096] Computer system 420 may also include one or more modules 452 that may, as described above and below, provide various functionality of the context-sensitive viewing system. For example, one module 452 may comprise the client-side context viewing module 410 of FIG. 4A, and may implement a graphical user interface on the client device 402. Module(s) 452 may be stored in a mass storage device as executable software codes that are executed by the computing device(s). This and other modules may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.
[0097] In general, the word "module," as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, Lua, C or C++. A software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, or Python.
It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices may be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, magnetic disc, or any other tangible medium, or as a digital download (and may be originally stored in a compressed or installable format that requires installation, decompression or decryption prior to execution). Such software code may be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules or computing device functionality described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage [0098] Computer system 420 may implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 420 to be a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 420 in response to processor(s) 424 executing one or more sequences of one or more instructions contained in main memory 426. Such instructions may be read into main memory 426 from another storage medium, such as storage device 430.
Execution of the sequences of instructions contained in main memory 426 causes processor(s) 424 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices may be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, magnetic disc, or any other tangible medium, or as a digital download (and may be originally stored in a compressed or installable format that requires installation, decompression or decryption prior to execution). Such software code may be stored, partially or fully, on a memory device of the executing computing device, for execution by the computing device. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules or computing device functionality described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage [0098] Computer system 420 may implement the techniques described herein using customized hard-wired logic, one or more ASICs or FPGAs, firmware and/or program logic which in combination with the computer system causes or programs computer system 420 to be a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 420 in response to processor(s) 424 executing one or more sequences of one or more instructions contained in main memory 426. Such instructions may be read into main memory 426 from another storage medium, such as storage device 430.
Execution of the sequences of instructions contained in main memory 426 causes processor(s) 424 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
[0099] The terms "non-transitory media," "computer-readable media,"
and similar terms, as used herein refers to any media that store data and/or instructions that cause a machine to operate in a specific fashion. Such non-transitory media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 430. Volatile media includes dynamic memory, such as main memory 426. Common forms of non-transitory media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge, and networked versions of the same.
and similar terms, as used herein refers to any media that store data and/or instructions that cause a machine to operate in a specific fashion. Such non-transitory media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 430. Volatile media includes dynamic memory, such as main memory 426. Common forms of non-transitory media include, for example, a floppy disk, a flexible disk, hard disk, solid state drive, magnetic tape, or any other magnetic data storage medium, a CD-ROM, any other optical data storage medium, any physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, NVRAM, any other memory chip or cartridge, and networked versions of the same.
[0100] Non-transitory media is distinct from but may be used in conjunction with transmission media. Transmission media participates in transferring information between non-transitory media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 422. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
[0101] Various forms of media may be involved in carrying one or more sequences of one or more instructions to processor 424 for execution. For example, the instructions may initially be carried on a magnetic disk or solid state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 420 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 422. Bus 422 carries the data to main memory 426, from which processor 424 retrieves and executes the instructions. The instructions received by main memory 426 may retrieve and execute the instructions. The instructions received by main memory 426 may optionally be stored on storage device 430 either before or after execution by processor 424.
[0102] Computer system 420 also includes a communication interface 438 coupled to bus 422. Communication interface 438 provides a two-way data communication coupling to a network link 440 that is connected to a local network 442. For example, communication interface 438 may be an integrated services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of telephone line.
As another example, communication interface 438 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN (or WAN component to communicate with a WAN). Wireless links may also be implemented. In any such implementation, communication interface 438 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
As another example, communication interface 438 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN (or WAN component to communicate with a WAN). Wireless links may also be implemented. In any such implementation, communication interface 438 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
[0103] Network link 440 typically provides data communication through one or more networks (for example, network 408 of FIG. 4A) to other data devices.
For example, network link 440 may provide a connection through local network to a host computer 444 or to data equipment operated by an Internet Service Provider (ISP) 446. ISP 446 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the "Internet" 448. Local network 442 and Internet 448 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks, on network link 440, and through communication interface 438, which carry the digital data to and from computer system 420, are example forms of transmission media.
For example, network link 440 may provide a connection through local network to a host computer 444 or to data equipment operated by an Internet Service Provider (ISP) 446. ISP 446 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the "Internet" 448. Local network 442 and Internet 448 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks, on network link 440, and through communication interface 438, which carry the digital data to and from computer system 420, are example forms of transmission media.
[0104] Computer system 420 can send messages and receive data, including program code, through the network(s), network link 440 and communication interface 438. In the Internet example, a server 450 might transmit a requested code for an application program through Internet 448, ISP 446, local network 442 and communication interface 438.
[0105] The received code may be executed by processor 424 as it is received, and/or stored in storage device 430, or other non-volatile storage for later execution.
[0106] In an embodiment, the context-sensitive viewing system is implemented by the computer system 420. For example, data objects, data visualization information, and/or contextual view information may be stored in the storage device 430, and/or in an external database accessible through the local network 442 (for example, database 406 of FIG. 4A). The user interfaces and/or operations of the context-sensitive viewing system may be implemented by modules 452 stored in the main memory 426, the ROM 428, and/or the storage device 430, and executed by the processor(s) 424.
[0107] The context-sensitive viewing system advantageously enables a user to view a particular set of data objects in multiple visualization contexts.
Previews of the set of data in other visualization may be quickly reviewed by the user to determine the most beneficial context for information extraction.
Further, manipulations by the user in one context are propagated to the other contexts, allowing fast analysis of the impacts of changes to the set of data.
Object Centric Data Model [0108] FIG. 5 illustrates an object-centric conceptual data model including an example database system 510 using an ontology 505. This description is provided for the purpose of providing an example and is not intended to limit the techniques to the example data model, the example database system, or the example database system's use of an ontology to represent information.
Previews of the set of data in other visualization may be quickly reviewed by the user to determine the most beneficial context for information extraction.
Further, manipulations by the user in one context are propagated to the other contexts, allowing fast analysis of the impacts of changes to the set of data.
Object Centric Data Model [0108] FIG. 5 illustrates an object-centric conceptual data model including an example database system 510 using an ontology 505. This description is provided for the purpose of providing an example and is not intended to limit the techniques to the example data model, the example database system, or the example database system's use of an ontology to represent information.
[0109] In one embodiment, a body of data is conceptually structured according to an object-centric data model represented by ontology 505. The conceptual data model is independent of any particular database used for durably storing one or more database(s) 509 based on the ontology 505. For example, each object of the conceptual data model may correspond to one or more rows in a relational database or an entry in Lightweight Directory Access Protocol (LDAP) database, or any combination of one or more databases.
[0110] An ontology 505, as noted above, may include stored information providing a data model for storage of data in the database 509. The ontology may be defined by one or more object types, which may each be associated with one or more property types. At the highest level of abstraction, data object 501 is a container for information representing things in the world. For example, data object 501 can represent an entity such as a person, a place, an organization, a market instrument, or other noun. Data object 501 can represent an event that happens at a point in time or for a duration. Data object 501 can represent a document or other unstructured data source such as an e-mail message, a news report, or a written paper or article. Each data object 501 is associated with a unique identifier that uniquely identifies the data object within the database system.
[0111] Different types of data objects may have different property types.
For example, a "Person" data object might have an "Eye Color" property type and an "Event" data object might have a "Date" property type. Each property 503 as represented by data in the database system 510 may have a property type defined by the ontology 505 used by the database 505.
For example, a "Person" data object might have an "Eye Color" property type and an "Event" data object might have a "Date" property type. Each property 503 as represented by data in the database system 510 may have a property type defined by the ontology 505 used by the database 505.
[0112] Objects may be instantiated in the database 509 in accordance with the corresponding object definition for the particular object in the ontology 505.
For example, a specific monetary payment (e.g., an object of type "event") of US$30.00 (e.g., a property of type "currency") taking place on 3/27/2009 (e.g., a property of type "date") may be stored in the database 509 as an event object with associated currency and date properties as defined within the ontology 505.
For example, a specific monetary payment (e.g., an object of type "event") of US$30.00 (e.g., a property of type "currency") taking place on 3/27/2009 (e.g., a property of type "date") may be stored in the database 509 as an event object with associated currency and date properties as defined within the ontology 505.
[0113] The data objects defined in the ontology 505 may support property multiplicity. In particular, a data object 501 may be allowed to have more than one property 503 of the same property type. For example, a "Person" data object might have multiple "Address" properties or multiple "Name" properties.
[0114] Each link 502 represents a connection between two data objects 501. In one embodiment, the connection is either through a relationship, an event, or through matching properties. A relationship connection may be asymmetrical or symmetrical. For example, "Person" data object A may be connected to "Person" data object B by a "Child Of" relationship (where "Person"
data object B has an asymmetric "Parent Of" relationship to "Person" data object A), a "Kin Of" symmetric relationship to "Person" data object C, and an asymmetric "Member Of" relationship to "Organization" data object X. The type of relationship between two data objects may vary depending on the types of the data objects.
For example, "Person" data object A may have an "Appears In" relationship with "Document" data object Y or have a "Participate In" relationship with "Event"
data object E. As an example of an event connection, two "Person" data objects may be connected by an "Airline Flight" data object representing a particular airline flight if they traveled together on that flight, or by a "Meeting" data object representing a particular meeting if they both attended that meeting. In one embodiment, when two data objects are connected by an event, they are also connected by relationships, in which each data object has a specific relationship to the event, such as, for example, an "Appears In" relationship.
data object B has an asymmetric "Parent Of" relationship to "Person" data object A), a "Kin Of" symmetric relationship to "Person" data object C, and an asymmetric "Member Of" relationship to "Organization" data object X. The type of relationship between two data objects may vary depending on the types of the data objects.
For example, "Person" data object A may have an "Appears In" relationship with "Document" data object Y or have a "Participate In" relationship with "Event"
data object E. As an example of an event connection, two "Person" data objects may be connected by an "Airline Flight" data object representing a particular airline flight if they traveled together on that flight, or by a "Meeting" data object representing a particular meeting if they both attended that meeting. In one embodiment, when two data objects are connected by an event, they are also connected by relationships, in which each data object has a specific relationship to the event, such as, for example, an "Appears In" relationship.
[0115] As an example of a matching properties connection, two "Person"
data objects representing a brother and a sister, may both have an "Address"
property that indicates where they live. If the brother and the sister live in the same home, then their "Address" properties likely contain similar, if not identical property values. In one embodiment, a link between two data objects may be established based on similar or matching properties (e.g., property types and/or property values) of the data objects. These are just some examples of the types of connections that may be represented by a link and other types of connections may be represented;
embodiments are not limited to any particular types of connections between data objects. For example, a document might contain references to two different objects.
For example, a document may contain a reference to a payment (one object), and a person (a second object). A link between these two objects may represent a connection between these two entities through their co-occurrence within the same document.
data objects representing a brother and a sister, may both have an "Address"
property that indicates where they live. If the brother and the sister live in the same home, then their "Address" properties likely contain similar, if not identical property values. In one embodiment, a link between two data objects may be established based on similar or matching properties (e.g., property types and/or property values) of the data objects. These are just some examples of the types of connections that may be represented by a link and other types of connections may be represented;
embodiments are not limited to any particular types of connections between data objects. For example, a document might contain references to two different objects.
For example, a document may contain a reference to a payment (one object), and a person (a second object). A link between these two objects may represent a connection between these two entities through their co-occurrence within the same document.
[0116] Each data object 501 can have multiple links with another data object 501 to form a link set 504. For example, two "Person" data objects representing a husband and a wife could be linked through a "Spouse Of"
relationship, a matching "Address" property, and one or more matching "Event"
properties (e.g., a wedding). Each link 502 as represented by data in a database may have a link type defined by the database ontology used by the database.
relationship, a matching "Address" property, and one or more matching "Event"
properties (e.g., a wedding). Each link 502 as represented by data in a database may have a link type defined by the database ontology used by the database.
[0117] FIG. 6 is a block diagram illustrating exemplary components and data that may be used in identifying and storing data according to an ontology. In this example, the ontology may be configured, and data in the data model populated, by a system of parsers and ontology configuration tools. In the embodiment of FIG. 6, input data 600 is provided to parser 602. The input data may comprise data from one or more sources. For example, an institution may have one or more databases with information on credit card transactions, rental cars, and people. The databases may contain a variety of related information and attributes about each type of data, such as a "date" for a credit card transaction, an address for a person, and a date for when a rental car is rented. The parser 602 is able to read a variety of source input data types and determine which type of data it is reading.
[0118] In accordance with the discussion above, the example ontology 505 comprises stored information providing the data model of data stored in database 509, and the ontology is defined by one or more object types 610, one or more property types 616, and one or more link types 630. Based on information determined by the parser 602 or other mapping of source input information to object type, one or more data objects 501 may be instantiated in the database 509 based on respective determined object types 610, and each of the objects 501 has one or more properties 503 that are instantiated based on property types 616. Two data objects 501 may be connected by one or more links 502 that may be instantiated based on link types 630. The property types 616 each may comprise one or more data types 618, such as a string, number, etc. Property types 616 may be instantiated based on a base property type 620. For example, a base property type 620 may be "Locations" and a property type 616 may be "Home."
[0119] In an embodiment, a user of the system uses an object type editor 624 to create and/or modify the object types 610 and define attributes of the object types. In an embodiment, a user of the system uses a property type editor 626 to create and/or modify the property types 616 and define attributes of the property types. In an embodiment, a user of the system uses link type editor 628 to create the link types 630. Alternatively, other programs, processes, or programmatic controls may be used to create link types and property types and define attributes, and using editors is not required.
[0120] In an embodiment, creating a property type 616 using the property type editor 626 involves defining at least one parser definition using a parser editor 622. A parser definition comprises metadata that informs parser 602 how to parse input data 600 to determine whether values in the input data can be assigned to the property type 616 that is associated with the parser definition. In an embodiment, each parser definition may comprise a regular expression parser or a code module parser 604B. In other embodiments, other kinds of parser definitions may be provided using scripts or other programmatic elements. Once defined, both a regular expression parser 604A and a code module parser 604B
can provide input to parser 602 to control parsing of input data 600.
can provide input to parser 602 to control parsing of input data 600.
[0121] Using the data types defined in the ontology, input data 600 may be parsed by the parser 602 determine which object type 610 should receive data from a record created from the input data, and which property types 616 should be assigned to data from individual field values in the input data. Based on the object-property mapping 601, the parser 602 selects one of the parser definitions that is associated with a property type in the input data. The parser parses an input data field using the selected parser definition, resulting in creating new or modified data 603. The new or modified data 603 is added to the database 509 according to ontology 505 by storing values of the new or modified data in a property of the specified property type. As a result, input data 600 having varying format or syntax can be created in database 509. The ontology 505 may be modified at any time using object type editor 624, property type editor 626, and link type editor 628, or under program control without human use of an editor. Parser editor 622 enables creating multiple parser definitions that can successfully parse input data 600 having varying format or syntax and determine which property types should be used to transform input data 600 into new or modified input data 603.
[0122] The properties, objects, and links (e.g. relationships) between the objects can be visualized using a graphical user interface (GUI). For example, as described above, FIG. 1A displays a user interface showing a graph representation of relationships between the data objects that are represented as nodes.
Further, as described above, relationships between data objects may be stored as links, or in some embodiments, as properties, where a relationship may be detected between the properties. In some cases, as stated above, the links may be directional.
For example, a payment link may have a direction associated with the payment, where one person object is a receiver of a payment, and another person object is the payer of payment.
Additional Implementation Details [0123] Each of the processes, methods, and algorithms described in the preceding sections may be embodied in, and fully or partially automated by, code modules executed by one or more computer systems or computer processors comprising computer hardware. The processes and algorithms may be implemented partially or wholly in application-specific circuitry.
Further, as described above, relationships between data objects may be stored as links, or in some embodiments, as properties, where a relationship may be detected between the properties. In some cases, as stated above, the links may be directional.
For example, a payment link may have a direction associated with the payment, where one person object is a receiver of a payment, and another person object is the payer of payment.
Additional Implementation Details [0123] Each of the processes, methods, and algorithms described in the preceding sections may be embodied in, and fully or partially automated by, code modules executed by one or more computer systems or computer processors comprising computer hardware. The processes and algorithms may be implemented partially or wholly in application-specific circuitry.
[0124] The various features and processes described above may be used independently of one another, or may be combined in various ways. All possible combinations and subcombinations are intended to fall within the scope of this disclosure. In addition, certain method or process blocks may be omitted in some implementations. The methods and processes described herein are also not limited to any particular sequence, and the blocks or states relating thereto can be performed in other sequences that are appropriate. For example, described blocks or states may be performed in an order other than that specifically disclosed, or multiple blocks or states may be combined in a single block or state. The example blocks or states may be performed in serial, in parallel, or in some other manner.
Blocks or states may be added to or removed from the disclosed example embodiments. The example systems and components described herein may be configured differently than described. For example, elements may be added to, removed from, or rearranged compared to the disclosed example embodiments.
Blocks or states may be added to or removed from the disclosed example embodiments. The example systems and components described herein may be configured differently than described. For example, elements may be added to, removed from, or rearranged compared to the disclosed example embodiments.
[0125] Conditional language, such as, among others, "can," "could,"
"might," or "may," unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.
"might," or "may," unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.
[0126] The term "comprising" as used herein should be given an inclusive rather than exclusive interpretation. For example, a general purpose computer comprising one or more processors should not be interpreted as excluding other computer components, and may possibly include such components as memory, input/output devices, and/or network interfaces, among others.
[0127] Any process descriptions, elements, or blocks in the flow diagrams described herein and/or depicted in the attached figures should be understood as potentially representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps in the process. Alternate implementations are included within the scope of the embodiments described herein in which elements or functions may be deleted, executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those skilled in the art.
[0128] It should be emphasized that many variations and modifications may be made to the above-described embodiments, the elements of which are to be understood as being among other acceptable examples. All such modifications and variations are intended to be included herein within the scope of this disclosure. The foregoing description details certain embodiments of the invention. It will be appreciated, however, that no matter how detailed the foregoing appears in text, the invention can be practiced in many ways. As is also stated above, it should be noted that the use of particular terminology when describing certain features or aspects of the invention should not be taken to imply that the terminology is being re-defined herein to be restricted to including any specific characteristics of the features or aspects of the invention with which that terminology is associated. The scope of the invention should therefore be construed in accordance with the appended claims and any equivalents thereof.
Claims (40)
1. A computer system comprising:
one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least:
an electronic database configured to store a plurality of data objects and properties associated with each of the data objects; and a context viewing module configured to:
generate a primary contextual view including a visualization of a set of data objects and associated properties;
generate one or more secondary contextual views, each secondary contextual view including respective secondary visualizations of the set of data objects;
receive a user input modifying the primary contextual view; and in response to receiving a user input modifying the primary contextual view, modify one or more of the secondary contextual views based at least in part on the user input.
one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least:
an electronic database configured to store a plurality of data objects and properties associated with each of the data objects; and a context viewing module configured to:
generate a primary contextual view including a visualization of a set of data objects and associated properties;
generate one or more secondary contextual views, each secondary contextual view including respective secondary visualizations of the set of data objects;
receive a user input modifying the primary contextual view; and in response to receiving a user input modifying the primary contextual view, modify one or more of the secondary contextual views based at least in part on the user input.
2. The computer system of Claim 1, wherein the context viewing module is further configured to:
in response to receiving a user input modifying the primary contextual view, determine which of the one or more of the secondary contextual views to modify.
in response to receiving a user input modifying the primary contextual view, determine which of the one or more of the secondary contextual views to modify.
3. The computer system of Claim 2, wherein modifying one or more of the secondary contextual views comprises modifying all of the one or more secondary contextual view.
4. The computer system of Claim 2, wherein modifying one or more of the secondary contextual views comprises modifying any of the one or more secondary contextual views that are currently viewable by a user.
5. The computer system of Claim 2, wherein modifying one or more of the secondary contextual views comprises modifying any of the one or more secondary contextual views that are immediately adjacent to the primary contextual view.
6. The computer system of Claim 1, wherein each of the visualization and/or the secondary visualizations include at least one of a graph, a map, a table, a timeline, a histogram, a list, a reader interface, or a postboard interface.
7. The computer system of Claim 1, wherein the one or more secondary contextual views comprise contextual previews.
8. The computer system of Claim 7, wherein the one or more secondary contextual views are viewable in a drawer or scrollbar on a user interface, wherein the one or more secondary contextual views are selectable by a user.
9. The computer system of Claim 1, wherein the one or more secondary contextual views are substantially the same size as the primary contextual view.
10. The computer system of Claim 9, wherein the one or more secondary contextual views are configured to be accessible by a user through the use of a scrollbar.
11. The computer system of Claim 10, wherein the scrollbar includes at least one of tick marks indicating the locations of the one or more secondary views or contextual previews accessible in pop up windows.
12. The computer system of Claim 9, wherein the one or more secondary contextual views are positioned laterally to the primary contextual view, and wherein the secondary contextual views are accessible by a user through a user input including at least one of a mouse cursor or a touch input.
13. The computer system of Claim 1, wherein a user input modifying the primary contextual view comprises at least one of adding data objects, removing data objects, modifying data objects, moving data objects, modifying properties associated with data objects, or modifying and/or manipulating links between data objects.
14. A computer system comprising:
one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least:
a context viewing module configured to:
display a first data visualization of a set of data objects and properties associated with data objects of the set of data objects;
provide one or more secondary data visualizations of the set of data objects;
receive a user input modifying the first data visualization;
and in response to receiving a user input modifying the first data visualization, implement modifications to at least some of the one or more secondary data visualizations based at least in part on the user input.
one or more hardware processors in communication with a computer readable medium storing software modules including instructions that are executable by the one or more hardware processors, the software modules including at least:
a context viewing module configured to:
display a first data visualization of a set of data objects and properties associated with data objects of the set of data objects;
provide one or more secondary data visualizations of the set of data objects;
receive a user input modifying the first data visualization;
and in response to receiving a user input modifying the first data visualization, implement modifications to at least some of the one or more secondary data visualizations based at least in part on the user input.
15. The computer system of Claim 14, wherein the context viewing module is further configured to:
in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are currently displayed to the user, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are currently displayed to the user, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
16. The computer system of Claim 14, wherein the context viewing module is further configured to:
in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are adjacent to the first data visualization, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
in response to receiving a user input modifying the modifying the first data visualization, determine which of the one or more of the secondary data visualizations are adjacent to the first data visualization, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
17. A computer-implemented method of updating multiple contextual views, the method comprising:
providing an electronic database configured to store a plurality of data objects and metadata associated with each of the plurality of data objects;
generating, by a computing system having one or more computer processors, based at least in part on the plurality of data objects and associated metadata, a primary contextual view and one or more secondary contextual views;
receiving, via an input device of the computing system, a user input modifying the primary contextual view;
determining, by the computing system, based on the received user input, modifications of the one or more secondary contextual views that correspond to the modification of the primary contextual view;
modifying at least some of the one or more secondary contextual views based on the determined modifications.
providing an electronic database configured to store a plurality of data objects and metadata associated with each of the plurality of data objects;
generating, by a computing system having one or more computer processors, based at least in part on the plurality of data objects and associated metadata, a primary contextual view and one or more secondary contextual views;
receiving, via an input device of the computing system, a user input modifying the primary contextual view;
determining, by the computing system, based on the received user input, modifications of the one or more secondary contextual views that correspond to the modification of the primary contextual view;
modifying at least some of the one or more secondary contextual views based on the determined modifications.
18. The computer-implemented method of Claim 17, further comprising:
providing, on an electronic display of the computing system, the generated primary contextual view and one or more of the secondary contextual views in a preview drawer.
providing, on an electronic display of the computing system, the generated primary contextual view and one or more of the secondary contextual views in a preview drawer.
19. The computer-implemented method of Claim 17, further comprising:
providing, on an electronic display of the computing system, the generated primary contextual view; and providing, on the electronic display of the computer system, a scrollbar that enables a user to scroll to any of the one or more of the secondary contextual views and view any of the one or more of the secondary contextual views on the electronic display.
providing, on an electronic display of the computing system, the generated primary contextual view; and providing, on the electronic display of the computer system, a scrollbar that enables a user to scroll to any of the one or more of the secondary contextual views and view any of the one or more of the secondary contextual views on the electronic display.
20. The computer-implemented method of Claim 17, wherein modifying at least some of the one or more secondary contextual views based on the determined modifications comprises modifying any secondary contextual views that are immediately viewable by a user.
21. A computer system comprising:
an electronic database configured to store a plurality of data objects and properties associated with each of the data objects; and one or more hardware processors in communication with a computer readable medium storing software instructions that are executable by the one or more hardware processors in order to cause the computer system to:
generate a primary contextual view including a primary visualization of a set of data objects and associated properties;
generate one or more secondary contextual views distinct from the primary contextual view, each secondary contextual view including respective secondary visualizations of the set of data objects, each of the secondary visualizations distinct from the primary visualization; and in response to receiving a user input modifying the primary contextual view, modify one or more of the secondary contextual views based at least in part on the user input.
an electronic database configured to store a plurality of data objects and properties associated with each of the data objects; and one or more hardware processors in communication with a computer readable medium storing software instructions that are executable by the one or more hardware processors in order to cause the computer system to:
generate a primary contextual view including a primary visualization of a set of data objects and associated properties;
generate one or more secondary contextual views distinct from the primary contextual view, each secondary contextual view including respective secondary visualizations of the set of data objects, each of the secondary visualizations distinct from the primary visualization; and in response to receiving a user input modifying the primary contextual view, modify one or more of the secondary contextual views based at least in part on the user input.
22. The computer system of Claim 21, wherein the software instructions are executable by the one or more hardware processors in order to further cause the computer system to:
further in response to receiving the user input modifying the primary contextual view, determine which of the one or more of the secondary contextual views to modify.
further in response to receiving the user input modifying the primary contextual view, determine which of the one or more of the secondary contextual views to modify.
23. The computer system of Claim 22, wherein modifying one or more of the secondary contextual views comprises modifying all of the one or more secondary contextual view.
24. The computer system of Claim 22, wherein modifying one or more of the secondary contextual views comprises modifying any of the one or more secondary contextual views that are currently viewable by a user.
25. The computer system of Claim 22, wherein modifying one or more of the secondary contextual views comprises modifying any of the one or more secondary contextual views that are immediately adjacent to the primary contextual view.
26. The computer system of Claim 21, wherein each of the primary visualization and/or the secondary visualizations include at least one of a graph, a map, a table, a timeline, a histogram, a list, a reader interface, or a postboard interface.
27. The computer system of Claim 21, wherein the one or more secondary contextual views comprise contextual previews.
28. The computer system of Claim 27, wherein the one or more secondary contextual views are viewable in a drawer or scrollbar on a user interface, wherein the one or more secondary contextual views are selectable by a user.
29. The computer system of Claim 21, wherein the one or more secondary contextual views are substantially the same size as the primary contextual view.
30. The computer system of Claim 29, wherein the one or more secondary contextual views are configured to be accessible by a user through the use of a scrollbar.
31. The computer system of Claim 30, wherein the scrollbar includes at least one of tick marks indicating the locations of the one or more secondary views or contextual previews accessible in pop up windows.
32. The computer system of Claim 29, wherein the one or more secondary contextual views are positioned laterally to the primary contextual view, and wherein the secondary contextual views are accessible by a user through a user input including at least one of a mouse cursor or a touch input.
33. The computer system of Claim 21, wherein a user input modifying the primary contextual view comprises at least one of adding data objects, removing data objects, modifying data objects, moving data objects, modifying properties associated with data objects, or modifying and/or manipulating links between data objects.
34. A computer system comprising:
one or more hardware processors in communication with a computer readable medium storing software instructions that are executable by the one or more hardware processors in order to cause the computer system to:
display a first data visualization of a set of data objects and properties associated with data objects of the set of data objects;
provide one or more secondary data visualizations of the set of data objects, the one or more secondary data visualizations distinct from the first data visualization; and in response to receiving a user input modifying the first data visualization, implement modifications to at least some of the one or more secondary data visualizations based at least in part on the user input.
one or more hardware processors in communication with a computer readable medium storing software instructions that are executable by the one or more hardware processors in order to cause the computer system to:
display a first data visualization of a set of data objects and properties associated with data objects of the set of data objects;
provide one or more secondary data visualizations of the set of data objects, the one or more secondary data visualizations distinct from the first data visualization; and in response to receiving a user input modifying the first data visualization, implement modifications to at least some of the one or more secondary data visualizations based at least in part on the user input.
35. The computer system of Claim 34, wherein further in response to receiving the user input modifying the first data visualization the software instructions are executable by the one or more hardware processors in order to cause the computer system to:
determine which of the one or more of the secondary data visualizations are currently displayed to the user; and implement modifications to the determined secondary data visualizations based at least in part on the user input.
determine which of the one or more of the secondary data visualizations are currently displayed to the user; and implement modifications to the determined secondary data visualizations based at least in part on the user input.
36. The computer system of Claim 34, wherein further in response to receiving the user input modifying the first data visualization the software instructions are executable by the one or more hardware processors in order to cause the computer system to:
determine which of the one or more of the secondary data visualizations are adjacent to the first data visualization, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
determine which of the one or more of the secondary data visualizations are adjacent to the first data visualization, and implement modifications to the determined secondary data visualizations based at least in part on the user input.
37. A computer-implemented method of updating multiple contextual views, the method comprising:
providing an electronic database configured to store a plurality of data objects and metadata associated with each of the plurality of data objects;
generating, by a computing system having one or more computer processors, based at least in part on the plurality of data objects and associated metadata, a primary contextual view and one or more secondary contextual views, the one or more secondary contextual views distinct from the primary contextual view;
receiving, via an input device of the computing system, a user input modifying the primary contextual view;
determining, by the computing system, based on the received user input, modifications of the one or more secondary contextual views that correspond to the modification of the primary contextual view;
modifying at least some of the one or more secondary contextual views based on the determined modifications.
providing an electronic database configured to store a plurality of data objects and metadata associated with each of the plurality of data objects;
generating, by a computing system having one or more computer processors, based at least in part on the plurality of data objects and associated metadata, a primary contextual view and one or more secondary contextual views, the one or more secondary contextual views distinct from the primary contextual view;
receiving, via an input device of the computing system, a user input modifying the primary contextual view;
determining, by the computing system, based on the received user input, modifications of the one or more secondary contextual views that correspond to the modification of the primary contextual view;
modifying at least some of the one or more secondary contextual views based on the determined modifications.
38. The computer-implemented method of Claim 37, further comprising:
providing, on an electronic display of the computing system, the generated primary contextual view and one or more of the secondary contextual views in a preview drawer.
providing, on an electronic display of the computing system, the generated primary contextual view and one or more of the secondary contextual views in a preview drawer.
39. The computer-implemented method of Claim 37, further comprising:
providing, on an electronic display of the computing system, the generated primary contextual view; and providing, on the electronic display of the computer system, a scrollbar that enables a user to scroll to any of the one or more of the secondary contextual views and view any of the one or more of the secondary contextual views on the electronic display.
providing, on an electronic display of the computing system, the generated primary contextual view; and providing, on the electronic display of the computer system, a scrollbar that enables a user to scroll to any of the one or more of the secondary contextual views and view any of the one or more of the secondary contextual views on the electronic display.
40. The computer-implemented method of Claim 37, wherein modifying at least some of the one or more secondary contextual views based on the determined modifications comprises modifying any secondary contextual views that are immediately viewable by a user.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361864048P | 2013-08-09 | 2013-08-09 | |
US61/864,048 | 2013-08-09 | ||
US14/095,798 | 2013-12-03 | ||
US14/095,798 US8713467B1 (en) | 2013-08-09 | 2013-12-03 | Context-sensitive views |
Publications (1)
Publication Number | Publication Date |
---|---|
CA2858586A1 true CA2858586A1 (en) | 2015-02-09 |
Family
ID=50514412
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA2858586A Abandoned CA2858586A1 (en) | 2013-08-09 | 2014-08-07 | Context-sensitive views |
Country Status (4)
Country | Link |
---|---|
US (4) | US8713467B1 (en) |
EP (1) | EP2838039A3 (en) |
AU (1) | AU2014210655B2 (en) |
CA (1) | CA2858586A1 (en) |
Families Citing this family (157)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8930331B2 (en) | 2007-02-21 | 2015-01-06 | Palantir Technologies | Providing unique views of data based on changes or rules |
US9348499B2 (en) | 2008-09-15 | 2016-05-24 | Palantir Technologies, Inc. | Sharing objects that rely on local resources with outside servers |
US8799240B2 (en) | 2011-06-23 | 2014-08-05 | Palantir Technologies, Inc. | System and method for investigating large amounts of data |
US9092482B2 (en) | 2013-03-14 | 2015-07-28 | Palantir Technologies, Inc. | Fair scheduling for mixed-query loads |
US9547693B1 (en) | 2011-06-23 | 2017-01-17 | Palantir Technologies Inc. | Periodic database search manager for multiple data sources |
US9792017B1 (en) | 2011-07-12 | 2017-10-17 | Domo, Inc. | Automatic creation of drill paths |
US10001898B1 (en) | 2011-07-12 | 2018-06-19 | Domo, Inc. | Automated provisioning of relational information for a summary data visualization |
US9202297B1 (en) | 2011-07-12 | 2015-12-01 | Domo, Inc. | Dynamic expansion of data visualizations |
US8732574B2 (en) | 2011-08-25 | 2014-05-20 | Palantir Technologies, Inc. | System and method for parameterizing documents for automatic workflow generation |
US8504542B2 (en) | 2011-09-02 | 2013-08-06 | Palantir Technologies, Inc. | Multi-row transactions |
US9348677B2 (en) | 2012-10-22 | 2016-05-24 | Palantir Technologies Inc. | System and method for batch evaluation programs |
US9123086B1 (en) | 2013-01-31 | 2015-09-01 | Palantir Technologies, Inc. | Automatically generating event objects from images |
US10037314B2 (en) | 2013-03-14 | 2018-07-31 | Palantir Technologies, Inc. | Mobile reports |
US10275778B1 (en) | 2013-03-15 | 2019-04-30 | Palantir Technologies Inc. | Systems and user interfaces for dynamic and interactive investigation based on automatic malfeasance clustering of related data in various data structures |
US8917274B2 (en) | 2013-03-15 | 2014-12-23 | Palantir Technologies Inc. | Event matrix based on integrated data |
US8909656B2 (en) | 2013-03-15 | 2014-12-09 | Palantir Technologies Inc. | Filter chains with associated multipath views for exploring large data sets |
US8788405B1 (en) | 2013-03-15 | 2014-07-22 | Palantir Technologies, Inc. | Generating data clusters with customizable analysis strategies |
US8937619B2 (en) | 2013-03-15 | 2015-01-20 | Palantir Technologies Inc. | Generating an object time series from data objects |
US8868486B2 (en) | 2013-03-15 | 2014-10-21 | Palantir Technologies Inc. | Time-sensitive cube |
US9965937B2 (en) | 2013-03-15 | 2018-05-08 | Palantir Technologies Inc. | External malware data item clustering and analysis |
US8799799B1 (en) | 2013-05-07 | 2014-08-05 | Palantir Technologies Inc. | Interactive geospatial map |
US9335897B2 (en) | 2013-08-08 | 2016-05-10 | Palantir Technologies Inc. | Long click display of a context menu |
US9223773B2 (en) | 2013-08-08 | 2015-12-29 | Palatir Technologies Inc. | Template system for custom document generation |
US8713467B1 (en) * | 2013-08-09 | 2014-04-29 | Palantir Technologies, Inc. | Context-sensitive views |
US9785317B2 (en) | 2013-09-24 | 2017-10-10 | Palantir Technologies Inc. | Presentation and analysis of user interaction data |
US8938686B1 (en) | 2013-10-03 | 2015-01-20 | Palantir Technologies Inc. | Systems and methods for analyzing performance of an entity |
US8812960B1 (en) | 2013-10-07 | 2014-08-19 | Palantir Technologies Inc. | Cohort-based presentation of user interaction data |
US9116975B2 (en) | 2013-10-18 | 2015-08-25 | Palantir Technologies Inc. | Systems and user interfaces for dynamic and interactive simultaneous querying of multiple data stores |
US8924872B1 (en) | 2013-10-18 | 2014-12-30 | Palantir Technologies Inc. | Overview user interface of emergency call data of a law enforcement agency |
US9021384B1 (en) | 2013-11-04 | 2015-04-28 | Palantir Technologies Inc. | Interactive vehicle information map |
US8868537B1 (en) | 2013-11-11 | 2014-10-21 | Palantir Technologies, Inc. | Simple web search |
US9105000B1 (en) | 2013-12-10 | 2015-08-11 | Palantir Technologies Inc. | Aggregating data from a plurality of data sources |
US9727622B2 (en) | 2013-12-16 | 2017-08-08 | Palantir Technologies, Inc. | Methods and systems for analyzing entity performance |
US9552615B2 (en) | 2013-12-20 | 2017-01-24 | Palantir Technologies Inc. | Automated database analysis to detect malfeasance |
US10356032B2 (en) | 2013-12-26 | 2019-07-16 | Palantir Technologies Inc. | System and method for detecting confidential information emails |
US8832832B1 (en) | 2014-01-03 | 2014-09-09 | Palantir Technologies Inc. | IP reputation |
US9043696B1 (en) | 2014-01-03 | 2015-05-26 | Palantir Technologies Inc. | Systems and methods for visual definition of data associations |
USD759052S1 (en) * | 2014-02-18 | 2016-06-14 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with graphical user interface |
US9483162B2 (en) | 2014-02-20 | 2016-11-01 | Palantir Technologies Inc. | Relationship visualizations |
US9009827B1 (en) | 2014-02-20 | 2015-04-14 | Palantir Technologies Inc. | Security sharing system |
US9727376B1 (en) | 2014-03-04 | 2017-08-08 | Palantir Technologies, Inc. | Mobile tasks |
US8935201B1 (en) | 2014-03-18 | 2015-01-13 | Palantir Technologies Inc. | Determining and extracting changed data from a data source |
US9836580B2 (en) | 2014-03-21 | 2017-12-05 | Palantir Technologies Inc. | Provider portal |
US9857958B2 (en) | 2014-04-28 | 2018-01-02 | Palantir Technologies Inc. | Systems and user interfaces for dynamic and interactive access of, investigation of, and analysis of data objects stored in one or more databases |
US9830388B2 (en) * | 2014-04-30 | 2017-11-28 | Excalibur Ip, Llc | Modular search object framework |
USD776140S1 (en) | 2014-04-30 | 2017-01-10 | Yahoo! Inc. | Display screen with graphical user interface for displaying search results as a stack of overlapping, actionable cards |
US9009171B1 (en) | 2014-05-02 | 2015-04-14 | Palantir Technologies Inc. | Systems and methods for active column filtering |
US9535974B1 (en) | 2014-06-30 | 2017-01-03 | Palantir Technologies Inc. | Systems and methods for identifying key phrase clusters within documents |
US9619557B2 (en) | 2014-06-30 | 2017-04-11 | Palantir Technologies, Inc. | Systems and methods for key phrase characterization of documents |
US9021260B1 (en) | 2014-07-03 | 2015-04-28 | Palantir Technologies Inc. | Malware data item analysis |
US9785773B2 (en) | 2014-07-03 | 2017-10-10 | Palantir Technologies Inc. | Malware data item analysis |
US9256664B2 (en) | 2014-07-03 | 2016-02-09 | Palantir Technologies Inc. | System and method for news events detection and visualization |
US9202249B1 (en) | 2014-07-03 | 2015-12-01 | Palantir Technologies Inc. | Data item clustering and analysis |
US10572496B1 (en) | 2014-07-03 | 2020-02-25 | Palantir Technologies Inc. | Distributed workflow system and database with access controls for city resiliency |
US9779150B1 (en) * | 2014-08-15 | 2017-10-03 | Tableau Software, Inc. | Systems and methods for filtering data used in data visualizations that use relationships |
US9710527B1 (en) | 2014-08-15 | 2017-07-18 | Tableau Software, Inc. | Systems and methods of arranging displayed elements in data visualizations and use relationships |
US9779147B1 (en) | 2014-08-15 | 2017-10-03 | Tableau Software, Inc. | Systems and methods to query and visualize data and relationships |
US9454281B2 (en) | 2014-09-03 | 2016-09-27 | Palantir Technologies Inc. | System for providing dynamic linked panels in user interface |
US9501851B2 (en) | 2014-10-03 | 2016-11-22 | Palantir Technologies Inc. | Time-series analysis system |
US9767172B2 (en) | 2014-10-03 | 2017-09-19 | Palantir Technologies Inc. | Data aggregation and analysis system |
US9785328B2 (en) | 2014-10-06 | 2017-10-10 | Palantir Technologies Inc. | Presentation of multivariate data on a graphical user interface of a computing system |
US9984133B2 (en) | 2014-10-16 | 2018-05-29 | Palantir Technologies Inc. | Schematic and database linking system |
US9229952B1 (en) | 2014-11-05 | 2016-01-05 | Palantir Technologies, Inc. | History preserving data pipeline system and method |
US9043894B1 (en) | 2014-11-06 | 2015-05-26 | Palantir Technologies Inc. | Malicious software detection in a computing system |
US9471452B2 (en) | 2014-12-01 | 2016-10-18 | Uptake Technologies, Inc. | Adaptive handling of operating data |
US9367872B1 (en) | 2014-12-22 | 2016-06-14 | Palantir Technologies Inc. | Systems and user interfaces for dynamic and interactive investigation of bad actor behavior based on automatic clustering of related data in various data structures |
US10362133B1 (en) | 2014-12-22 | 2019-07-23 | Palantir Technologies Inc. | Communication data processing architecture |
US10552994B2 (en) | 2014-12-22 | 2020-02-04 | Palantir Technologies Inc. | Systems and interactive user interfaces for dynamic retrieval, analysis, and triage of data items |
US9348920B1 (en) | 2014-12-22 | 2016-05-24 | Palantir Technologies Inc. | Concept indexing among database of documents using machine learning techniques |
US9870205B1 (en) | 2014-12-29 | 2018-01-16 | Palantir Technologies Inc. | Storing logical units of program code generated using a dynamic programming notebook user interface |
US9817563B1 (en) | 2014-12-29 | 2017-11-14 | Palantir Technologies Inc. | System and method of generating data points from one or more data stores of data items for chart creation and manipulation |
US9335911B1 (en) | 2014-12-29 | 2016-05-10 | Palantir Technologies Inc. | Interactive user interface for dynamic data analysis exploration and query processing |
US10372879B2 (en) | 2014-12-31 | 2019-08-06 | Palantir Technologies Inc. | Medical claims lead summary report generation |
US11302426B1 (en) | 2015-01-02 | 2022-04-12 | Palantir Technologies Inc. | Unified data interface and system |
US10387834B2 (en) | 2015-01-21 | 2019-08-20 | Palantir Technologies Inc. | Systems and methods for accessing and storing snapshots of a remote application in a document |
US9727560B2 (en) | 2015-02-25 | 2017-08-08 | Palantir Technologies Inc. | Systems and methods for organizing and identifying documents via hierarchies and dimensions of tags |
EP3070622A1 (en) | 2015-03-16 | 2016-09-21 | Palantir Technologies, Inc. | Interactive user interfaces for location-based data analysis |
US9886467B2 (en) | 2015-03-19 | 2018-02-06 | Plantir Technologies Inc. | System and method for comparing and visualizing data entities and data entity series |
US10254751B2 (en) | 2015-06-05 | 2019-04-09 | Uptake Technologies, Inc. | Local analytics at an asset |
US10176279B2 (en) | 2015-06-05 | 2019-01-08 | Uptake Technologies, Inc. | Dynamic execution of predictive models and workflows |
US10579750B2 (en) | 2015-06-05 | 2020-03-03 | Uptake Technologies, Inc. | Dynamic execution of predictive models |
US10628834B1 (en) | 2015-06-16 | 2020-04-21 | Palantir Technologies Inc. | Fraud lead detection system for efficiently processing database-stored data and automatically generating natural language explanatory information of system results for display in interactive user interfaces |
US10878385B2 (en) | 2015-06-19 | 2020-12-29 | Uptake Technologies, Inc. | Computer system and method for distributing execution of a predictive model |
AU2016296471B2 (en) | 2015-07-17 | 2021-10-28 | Crown Equipment Corporation | Processing device having a graphical user interface for industrial vehicle |
US9418337B1 (en) | 2015-07-21 | 2016-08-16 | Palantir Technologies Inc. | Systems and models for data analytics |
US9454785B1 (en) | 2015-07-30 | 2016-09-27 | Palantir Technologies Inc. | Systems and user interfaces for holistic, data-driven investigation of bad actor behavior based on clustering and scoring of related data |
US9996595B2 (en) | 2015-08-03 | 2018-06-12 | Palantir Technologies, Inc. | Providing full data provenance visualization for versioned datasets |
US9456000B1 (en) | 2015-08-06 | 2016-09-27 | Palantir Technologies Inc. | Systems, methods, user interfaces, and computer-readable media for investigating potential malicious communications |
US10489391B1 (en) | 2015-08-17 | 2019-11-26 | Palantir Technologies Inc. | Systems and methods for grouping and enriching data items accessed from one or more databases for presentation in a user interface |
US9600146B2 (en) | 2015-08-17 | 2017-03-21 | Palantir Technologies Inc. | Interactive geospatial map |
US10102369B2 (en) | 2015-08-19 | 2018-10-16 | Palantir Technologies Inc. | Checkout system executable code monitoring, and user account compromise determination system |
US10853378B1 (en) | 2015-08-25 | 2020-12-01 | Palantir Technologies Inc. | Electronic note management via a connected entity graph |
US11150917B2 (en) | 2015-08-26 | 2021-10-19 | Palantir Technologies Inc. | System for data aggregation and analysis of data from a plurality of data sources |
US9485265B1 (en) | 2015-08-28 | 2016-11-01 | Palantir Technologies Inc. | Malicious activity detection system capable of efficiently processing data accessed from databases and generating alerts for display in interactive user interfaces |
US10706434B1 (en) | 2015-09-01 | 2020-07-07 | Palantir Technologies Inc. | Methods and systems for determining location information |
US9576015B1 (en) | 2015-09-09 | 2017-02-21 | Palantir Technologies, Inc. | Domain-specific language for dataset transformations |
JP2018537747A (en) | 2015-09-17 | 2018-12-20 | アップテイク テクノロジーズ、インコーポレイテッド | Computer system and method for sharing asset-related information between data platforms over a network |
US10296617B1 (en) | 2015-10-05 | 2019-05-21 | Palantir Technologies Inc. | Searches of highly structured data |
US10623294B2 (en) | 2015-12-07 | 2020-04-14 | Uptake Technologies, Inc. | Local analytics device |
US9542446B1 (en) | 2015-12-17 | 2017-01-10 | Palantir Technologies, Inc. | Automatic generation of composite datasets based on hierarchical fields |
US9823818B1 (en) | 2015-12-29 | 2017-11-21 | Palantir Technologies Inc. | Systems and interactive user interfaces for automatic generation of temporal representation of data objects |
US10089289B2 (en) | 2015-12-29 | 2018-10-02 | Palantir Technologies Inc. | Real-time document annotation |
US9612723B1 (en) | 2015-12-30 | 2017-04-04 | Palantir Technologies Inc. | Composite graphical interface with shareable data-objects |
US11295217B2 (en) | 2016-01-14 | 2022-04-05 | Uptake Technologies, Inc. | Localized temporal model forecasting |
KR102159072B1 (en) * | 2016-03-08 | 2020-09-24 | 비즈리드 엘엘씨 | Systems and methods for content reinforcement and reading education and comprehension |
US10510006B2 (en) | 2016-03-09 | 2019-12-17 | Uptake Technologies, Inc. | Handling of predictive models based on asset location |
US10698938B2 (en) | 2016-03-18 | 2020-06-30 | Palantir Technologies Inc. | Systems and methods for organizing and identifying documents via hierarchies and dimensions of tags |
US10796235B2 (en) | 2016-03-25 | 2020-10-06 | Uptake Technologies, Inc. | Computer systems and methods for providing a visualization of asset event and signal data |
US10333775B2 (en) | 2016-06-03 | 2019-06-25 | Uptake Technologies, Inc. | Facilitating the provisioning of a local analytics device |
US10719188B2 (en) | 2016-07-21 | 2020-07-21 | Palantir Technologies Inc. | Cached database and synchronization system for providing dynamic linked panels in user interface |
US10324609B2 (en) | 2016-07-21 | 2019-06-18 | Palantir Technologies Inc. | System for providing dynamic linked panels in user interface |
US10437840B1 (en) | 2016-08-19 | 2019-10-08 | Palantir Technologies Inc. | Focused probabilistic entity resolution from multiple data sources |
US10210037B2 (en) | 2016-08-25 | 2019-02-19 | Uptake Technologies, Inc. | Interface tool for asset fault analysis |
CN106909365A (en) | 2016-08-30 | 2017-06-30 | 阿里巴巴集团控股有限公司 | The methods of exhibiting and device of information |
US10474932B2 (en) | 2016-09-01 | 2019-11-12 | Uptake Technologies, Inc. | Detection of anomalies in multivariate data |
USD841687S1 (en) * | 2016-09-22 | 2019-02-26 | Brainlab Ag | Display screen with an animated graphical user interface for medical software |
US10055385B2 (en) * | 2016-10-21 | 2018-08-21 | Business Objects Software Limited | Visual preview based on analytical metadata |
US10318630B1 (en) | 2016-11-21 | 2019-06-11 | Palantir Technologies Inc. | Analysis of large bodies of textual data |
US10936183B2 (en) | 2016-11-22 | 2021-03-02 | Crown Equipment Corporation | User interface device for industrial vehicle |
US10228925B2 (en) | 2016-12-19 | 2019-03-12 | Uptake Technologies, Inc. | Systems, devices, and methods for deploying one or more artifacts to a deployment environment |
US10460602B1 (en) | 2016-12-28 | 2019-10-29 | Palantir Technologies Inc. | Interactive vehicle information mapping system |
US10579961B2 (en) | 2017-01-26 | 2020-03-03 | Uptake Technologies, Inc. | Method and system of identifying environment features for use in analyzing asset operation |
US10671039B2 (en) | 2017-05-03 | 2020-06-02 | Uptake Technologies, Inc. | Computer system and method for predicting an abnormal event at a wind turbine in a cluster |
US10255526B2 (en) | 2017-06-09 | 2019-04-09 | Uptake Technologies, Inc. | Computer system and method for classifying temporal patterns of change in images of an area |
US10956406B2 (en) | 2017-06-12 | 2021-03-23 | Palantir Technologies Inc. | Propagated deletion of database records and derived data |
US10403011B1 (en) | 2017-07-18 | 2019-09-03 | Palantir Technologies Inc. | Passing system with an interactive user interface |
GB201716303D0 (en) * | 2017-10-05 | 2017-11-22 | Palantir Technologies Inc | System and method for querying a data repository |
GB201716304D0 (en) * | 2017-10-05 | 2017-11-22 | Palantir Technologies Inc | Data analysis system and method |
US11232371B2 (en) | 2017-10-19 | 2022-01-25 | Uptake Technologies, Inc. | Computer system and method for detecting anomalies in multivariate data |
US10552246B1 (en) | 2017-10-24 | 2020-02-04 | Uptake Technologies, Inc. | Computer system and method for handling non-communicative assets |
US10379982B2 (en) | 2017-10-31 | 2019-08-13 | Uptake Technologies, Inc. | Computer system and method for performing a virtual load test |
US10635519B1 (en) | 2017-11-30 | 2020-04-28 | Uptake Technologies, Inc. | Systems and methods for detecting and remedying software anomalies |
US10815966B1 (en) | 2018-02-01 | 2020-10-27 | Uptake Technologies, Inc. | Computer system and method for determining an orientation of a wind turbine nacelle |
US10169135B1 (en) | 2018-03-02 | 2019-01-01 | Uptake Technologies, Inc. | Computer system and method of detecting manufacturing network anomalies |
US10554518B1 (en) | 2018-03-02 | 2020-02-04 | Uptake Technologies, Inc. | Computer system and method for evaluating health of nodes in a manufacturing network |
US11599369B1 (en) | 2018-03-08 | 2023-03-07 | Palantir Technologies Inc. | Graphical user interface configuration system |
US10754822B1 (en) | 2018-04-18 | 2020-08-25 | Palantir Technologies Inc. | Systems and methods for ontology migration |
US10635095B2 (en) | 2018-04-24 | 2020-04-28 | Uptake Technologies, Inc. | Computer system and method for creating a supervised failure model |
US10885021B1 (en) | 2018-05-02 | 2021-01-05 | Palantir Technologies Inc. | Interactive interpreter and graphical user interface |
US10860599B2 (en) | 2018-06-11 | 2020-12-08 | Uptake Technologies, Inc. | Tool for creating and deploying configurable pipelines |
US11119630B1 (en) | 2018-06-19 | 2021-09-14 | Palantir Technologies Inc. | Artificial intelligence assisted evaluations and user interface for same |
USD923042S1 (en) * | 2018-07-04 | 2021-06-22 | The Fourth Paradigm (Beijing) Tech Co Ltd | Display screen with transitional graphical user interface |
US10579932B1 (en) | 2018-07-10 | 2020-03-03 | Uptake Technologies, Inc. | Computer system and method for creating and deploying an anomaly detection model based on streaming data |
US11119472B2 (en) | 2018-09-28 | 2021-09-14 | Uptake Technologies, Inc. | Computer system and method for evaluating an event prediction model |
US11181894B2 (en) | 2018-10-15 | 2021-11-23 | Uptake Technologies, Inc. | Computer system and method of defining a set of anomaly thresholds for an anomaly detection model |
US11257263B1 (en) | 2018-10-23 | 2022-02-22 | Palantir Technologies Inc. | Systems and methods for generating dynamic pipeline visualizations |
US11480934B2 (en) | 2019-01-24 | 2022-10-25 | Uptake Technologies, Inc. | Computer system and method for creating an event prediction model |
US11030067B2 (en) | 2019-01-29 | 2021-06-08 | Uptake Technologies, Inc. | Computer system and method for presenting asset insights at a graphical user interface |
US11797550B2 (en) | 2019-01-30 | 2023-10-24 | Uptake Technologies, Inc. | Data science platform |
US11208986B2 (en) | 2019-06-27 | 2021-12-28 | Uptake Technologies, Inc. | Computer system and method for detecting irregular yaw activity at a wind turbine |
US10975841B2 (en) | 2019-08-02 | 2021-04-13 | Uptake Technologies, Inc. | Computer system and method for detecting rotor imbalance at a wind turbine |
US11586771B2 (en) * | 2019-10-18 | 2023-02-21 | Glance, LLC | Integrated visualization of multiple app data |
US11283699B2 (en) | 2020-01-17 | 2022-03-22 | Vmware, Inc. | Practical overlay network latency measurement in datacenter |
US11892830B2 (en) | 2020-12-16 | 2024-02-06 | Uptake Technologies, Inc. | Risk assessment at power substations |
US11336533B1 (en) | 2021-01-08 | 2022-05-17 | Vmware, Inc. | Network visualization of correlations between logical elements and associated physical elements |
US11855862B2 (en) | 2021-09-17 | 2023-12-26 | Vmware, Inc. | Tagging packets for monitoring and analysis |
CN114397994A (en) * | 2021-12-20 | 2022-04-26 | 北京旷视科技有限公司 | Object management method, electronic equipment and storage medium |
Family Cites Families (419)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5109399A (en) | 1989-08-18 | 1992-04-28 | Alamo City Technologies, Inc. | Emergency call locating system |
US5632987A (en) | 1992-12-31 | 1997-05-27 | Mycogen Corporation | Bacillus thuringiensis toxins active against corn rootworm larvae |
FR2684214B1 (en) | 1991-11-22 | 1997-04-04 | Sepro Robotique | INDEXING CARD FOR GEOGRAPHIC INFORMATION SYSTEM AND SYSTEM INCLUDING APPLICATION. |
US5670987A (en) | 1993-09-21 | 1997-09-23 | Kabushiki Kaisha Toshiba | Virtual manipulating apparatus and method |
US6877137B1 (en) | 1998-04-09 | 2005-04-05 | Rose Blush Software Llc | System, method and computer program product for mediating notes and note sub-notes linked or otherwise associated with stored or networked web pages |
US5777549A (en) | 1995-03-29 | 1998-07-07 | Cabletron Systems, Inc. | Method and apparatus for policy-based alarm notification in a distributed network management environment |
US6366933B1 (en) | 1995-10-27 | 2002-04-02 | At&T Corp. | Method and apparatus for tracking and viewing changes on the web |
US5845300A (en) | 1996-06-05 | 1998-12-01 | Microsoft Corporation | Method and apparatus for suggesting completions for a partially entered data item based on previously-entered, associated data items |
US5798769A (en) | 1996-08-15 | 1998-08-25 | Xerox Corporation | Method and apparatus for maintaining links between graphic objects in a free-form graphics display system |
CA2187704C (en) | 1996-10-11 | 1999-05-04 | Darcy Kim Rossmo | Expert system method of performing crime site analysis |
US5974572A (en) | 1996-10-15 | 1999-10-26 | Mercury Interactive Corporation | Software system and methods for generating a load test using a server access log |
US5870559A (en) | 1996-10-15 | 1999-02-09 | Mercury Interactive | Software system and associated methods for facilitating the analysis and management of web sites |
US6026233A (en) | 1997-05-27 | 2000-02-15 | Microsoft Corporation | Method and apparatus for presenting and selecting options to modify a programming language statement |
US6091956A (en) | 1997-06-12 | 2000-07-18 | Hollenberg; Dennis D. | Situation information system |
JP3636272B2 (en) | 1998-02-09 | 2005-04-06 | 富士通株式会社 | Icon display method, apparatus thereof, and recording medium |
US6247019B1 (en) | 1998-03-17 | 2001-06-12 | Prc Public Sector, Inc. | Object-based geographic information system (GIS) |
US7168039B2 (en) | 1998-06-02 | 2007-01-23 | International Business Machines Corporation | Method and system for reducing the horizontal space required for displaying a column containing text data |
US6742003B2 (en) | 2001-04-30 | 2004-05-25 | Microsoft Corporation | Apparatus and accompanying methods for visualizing clusters of data and hierarchical cluster classifications |
US6577304B1 (en) | 1998-08-14 | 2003-06-10 | I2 Technologies Us, Inc. | System and method for visually representing a supply chain |
US6161098A (en) | 1998-09-14 | 2000-12-12 | Folio (Fn), Inc. | Method and apparatus for enabling small investors with a portfolio of securities to manage taxable events within the portfolio |
US6232971B1 (en) | 1998-09-23 | 2001-05-15 | International Business Machines Corporation | Variable modality child windows |
US6279018B1 (en) | 1998-12-21 | 2001-08-21 | Kudrollis Software Inventions Pvt. Ltd. | Abbreviating and compacting text to cope with display space constraint in computer software |
US6631496B1 (en) | 1999-03-22 | 2003-10-07 | Nec Corporation | System for personalizing, organizing and managing web information |
US6369835B1 (en) | 1999-05-18 | 2002-04-09 | Microsoft Corporation | Method and system for generating a movie file from a slide show presentation |
US6714936B1 (en) | 1999-05-25 | 2004-03-30 | Nevin, Iii Rocky Harry W. | Method and apparatus for displaying data stored in linked nodes |
US6307573B1 (en) | 1999-07-22 | 2001-10-23 | Barbara L. Barros | Graphic-information flow method and system for visually analyzing patterns and relationships |
US8397177B2 (en) * | 1999-07-22 | 2013-03-12 | Tavusi Data Solutions Llc | Graphic-information flow method and system for visually analyzing patterns and relationships |
US7039863B1 (en) | 1999-07-23 | 2006-05-02 | Adobe Systems Incorporated | Computer generation of documents using layout elements and content elements |
US7373592B2 (en) | 1999-07-30 | 2008-05-13 | Microsoft Corporation | Modeless child windows for application programs |
US6560620B1 (en) | 1999-08-03 | 2003-05-06 | Aplix Research, Inc. | Hierarchical document comparison system and method |
US6976210B1 (en) | 1999-08-31 | 2005-12-13 | Lucent Technologies Inc. | Method and apparatus for web-site-independent personalization from multiple sites having user-determined extraction functionality |
US20020174201A1 (en) | 1999-09-30 | 2002-11-21 | Ramer Jon E. | Dynamic configuration of context-sensitive personal sites and membership channels |
US7716077B1 (en) * | 1999-11-22 | 2010-05-11 | Accenture Global Services Gmbh | Scheduling and planning maintenance and service in a network-based supply chain environment |
FR2806183B1 (en) | 1999-12-01 | 2006-09-01 | Cartesis S A | DEVICE AND METHOD FOR INSTANT CONSOLIDATION, ENRICHMENT AND "REPORTING" OR BACKGROUND OF INFORMATION IN A MULTIDIMENSIONAL DATABASE |
US7194680B1 (en) | 1999-12-07 | 2007-03-20 | Adobe Systems Incorporated | Formatting content by example |
US6859909B1 (en) | 2000-03-07 | 2005-02-22 | Microsoft Corporation | System and method for annotating web-based documents |
US6456997B1 (en) | 2000-04-12 | 2002-09-24 | International Business Machines Corporation | System and method for dynamically generating an invisible hierarchy in a planning system |
JP4325075B2 (en) | 2000-04-21 | 2009-09-02 | ソニー株式会社 | Data object management device |
US6915289B1 (en) | 2000-05-04 | 2005-07-05 | International Business Machines Corporation | Using an index to access a subject multi-dimensional database |
US7269786B1 (en) | 2000-05-04 | 2007-09-11 | International Business Machines Corporation | Navigating an index to access a subject multi-dimensional database |
US6642945B1 (en) | 2000-05-04 | 2003-11-04 | Microsoft Corporation | Method and system for optimizing a visual display for handheld computer systems |
US6594672B1 (en) | 2000-06-01 | 2003-07-15 | Hyperion Solutions Corporation | Generating multidimensional output using meta-models and meta-outlines |
US6839745B1 (en) * | 2000-07-19 | 2005-01-04 | Verizon Corporate Services Group Inc. | System and method for generating reports in a telecommunication system |
US7278105B1 (en) | 2000-08-21 | 2007-10-02 | Vignette Corporation | Visualization and analysis of user clickpaths |
US20020065708A1 (en) | 2000-09-22 | 2002-05-30 | Hikmet Senay | Method and system for interactive visual analyses of organizational interactions |
AUPR033800A0 (en) | 2000-09-25 | 2000-10-19 | Telstra R & D Management Pty Ltd | A document categorisation system |
US6829621B2 (en) | 2000-10-06 | 2004-12-07 | International Business Machines Corporation | Automatic determination of OLAP cube dimensions |
US8117281B2 (en) | 2006-11-02 | 2012-02-14 | Addnclick, Inc. | Using internet content as a means to establish live social networks by linking internet users to each other who are simultaneously engaged in the same and/or similar content |
US8707185B2 (en) | 2000-10-10 | 2014-04-22 | Addnclick, Inc. | Dynamic information management system and method for content delivery and sharing in content-, metadata- and viewer-based, live social networking among users concurrently engaged in the same and/or similar content |
JP2002123530A (en) | 2000-10-12 | 2002-04-26 | Hitachi Ltd | Method and device for visualizing multidimensional data |
US6738770B2 (en) | 2000-11-04 | 2004-05-18 | Deep Sky Software, Inc. | System and method for filtering and sorting data |
US9053222B2 (en) | 2002-05-17 | 2015-06-09 | Lawrence A. Lynn | Patient safety processor |
US6516268B2 (en) | 2001-02-16 | 2003-02-04 | Wizeguides.Com Inc. | Bundled map guide |
US6985950B1 (en) | 2001-03-06 | 2006-01-10 | Microsoft Corporation | System for creating a space-efficient document categorizer for training and testing of automatic categorization engines |
US7043702B2 (en) | 2001-03-15 | 2006-05-09 | Xerox Corporation | Method for visualizing user path through a web site and a path's associated information scent |
US9256356B2 (en) | 2001-03-29 | 2016-02-09 | International Business Machines Corporation | Method and system for providing feedback for docking a content pane in a host window |
US6775675B1 (en) | 2001-04-04 | 2004-08-10 | Sagemetrics Corporation | Methods for abstracting data from various data structures and managing the presentation of the data |
KR20040004619A (en) * | 2001-05-11 | 2004-01-13 | 컴퓨터 어소시에이츠 싱크, 인코포레이티드 | Method and system for transforming legacy software applications into modern object-oriented systems |
US6980984B1 (en) | 2001-05-16 | 2005-12-27 | Kanisa, Inc. | Content provider systems and methods using structured data |
US7865427B2 (en) | 2001-05-30 | 2011-01-04 | Cybersource Corporation | Method and apparatus for evaluating fraud risk in an electronic commerce transaction |
US6828920B2 (en) | 2001-06-04 | 2004-12-07 | Lockheed Martin Orincon Corporation | System and method for classifying vehicles |
US8001465B2 (en) | 2001-06-26 | 2011-08-16 | Kudrollis Software Inventions Pvt. Ltd. | Compacting an information array display to cope with two dimensional display space constraint |
US20030039948A1 (en) | 2001-08-09 | 2003-02-27 | Donahue Steven J. | Voice enabled tutorial system and method |
EP1435058A4 (en) | 2001-10-11 | 2005-12-07 | Visualsciences Llc | System, method, and computer program product for processing and visualization of information |
US7611602B2 (en) | 2001-12-13 | 2009-11-03 | Urban Mapping, Llc | Method of producing maps and other objects configured for presentation of spatially-related layers of data |
US20070203771A1 (en) * | 2001-12-17 | 2007-08-30 | Caballero Richard J | System and method for processing complex orders |
US7970240B1 (en) | 2001-12-17 | 2011-06-28 | Google Inc. | Method and apparatus for archiving and visualizing digital images |
US7139800B2 (en) * | 2002-01-16 | 2006-11-21 | Xerox Corporation | User interface for a message-based system having embedded information management capabilities |
US7454466B2 (en) * | 2002-01-16 | 2008-11-18 | Xerox Corporation | Method and system for flexible workflow management |
US7640173B2 (en) | 2002-01-17 | 2009-12-29 | Applied Medical Software, Inc. | Method and system for evaluating a physician's economic performance and gainsharing of physician services |
US7546245B2 (en) | 2002-01-17 | 2009-06-09 | Amsapplied Medical Software, Inc. | Method and system for gainsharing of physician services |
US7162475B2 (en) | 2002-04-17 | 2007-01-09 | Ackerman David M | Method for user verification and authentication and multimedia processing for interactive database management and method for viewing the multimedia |
US7171427B2 (en) | 2002-04-26 | 2007-01-30 | Oracle International Corporation | Methods of navigating a cube that is implemented as a relational object |
US20040126840A1 (en) | 2002-12-23 | 2004-07-01 | Affymetrix, Inc. | Method, system and computer software for providing genomic ontological data |
US20040012633A1 (en) | 2002-04-26 | 2004-01-22 | Affymetrix, Inc., A Corporation Organized Under The Laws Of Delaware | System, method, and computer program product for dynamic display, and analysis of biological sequence data |
US7703021B1 (en) | 2002-05-24 | 2010-04-20 | Sparta Systems, Inc. | Defining user access in highly-configurable systems |
JP2003345810A (en) | 2002-05-28 | 2003-12-05 | Hitachi Ltd | Method and system for document retrieval and document retrieval result display system |
US20030229848A1 (en) | 2002-06-05 | 2003-12-11 | Udo Arend | Table filtering in a computer user interface |
US7103854B2 (en) | 2002-06-27 | 2006-09-05 | Tele Atlas North America, Inc. | System and method for associating text and graphical views of map information |
CA2398103A1 (en) | 2002-08-14 | 2004-02-14 | March Networks Corporation | Multi-dimensional table filtering system |
US7127352B2 (en) | 2002-09-30 | 2006-10-24 | Lucent Technologies Inc. | System and method for providing accurate local maps for a central service |
WO2004036461A2 (en) | 2002-10-14 | 2004-04-29 | Battelle Memorial Institute | Information reservoir |
US20040143602A1 (en) | 2002-10-18 | 2004-07-22 | Antonio Ruiz | Apparatus, system and method for automated and adaptive digital image/video surveillance for events and configurations using a rich multimedia relational database |
US20040085318A1 (en) | 2002-10-31 | 2004-05-06 | Philipp Hassler | Graphics generation and integration |
US8589273B2 (en) | 2002-12-23 | 2013-11-19 | Ge Corporate Financial Services, Inc. | Methods and systems for managing risk management information |
US7627552B2 (en) | 2003-03-27 | 2009-12-01 | Microsoft Corporation | System and method for filtering and organizing items based on common elements |
US7280038B2 (en) | 2003-04-09 | 2007-10-09 | John Robinson | Emergency response data transmission system |
KR100996029B1 (en) | 2003-04-29 | 2010-11-22 | 삼성전자주식회사 | Apparatus and method for coding of low density parity check code |
US9607092B2 (en) | 2003-05-20 | 2017-03-28 | Excalibur Ip, Llc | Mapping method and system |
US20050027705A1 (en) | 2003-05-20 | 2005-02-03 | Pasha Sadri | Mapping method and system |
US7620648B2 (en) | 2003-06-20 | 2009-11-17 | International Business Machines Corporation | Universal annotation configuration and deployment |
US7055110B2 (en) | 2003-07-28 | 2006-05-30 | Sig G Kupka | Common on-screen zone for menu activation and stroke input |
US7363581B2 (en) | 2003-08-12 | 2008-04-22 | Accenture Global Services Gmbh | Presentation generator |
US7334195B2 (en) | 2003-10-14 | 2008-02-19 | Microsoft Corporation | System and process for presenting search results in a histogram/cluster format |
US7584172B2 (en) | 2003-10-16 | 2009-09-01 | Sap Ag | Control for selecting data query and visual configuration |
US20050125715A1 (en) | 2003-12-04 | 2005-06-09 | Fabrizio Di Franco | Method of saving data in a graphical user interface |
US7818658B2 (en) | 2003-12-09 | 2010-10-19 | Yi-Chih Chen | Multimedia presentation system |
US7917376B2 (en) | 2003-12-29 | 2011-03-29 | Montefiore Medical Center | System and method for monitoring patient care |
US7872669B2 (en) | 2004-01-22 | 2011-01-18 | Massachusetts Institute Of Technology | Photo-based mobile deixis system and related techniques |
US7343552B2 (en) | 2004-02-12 | 2008-03-11 | Fuji Xerox Co., Ltd. | Systems and methods for freeform annotations |
US20050180330A1 (en) | 2004-02-17 | 2005-08-18 | Touchgraph Llc | Method of animating transitions and stabilizing node motion during dynamic graph navigation |
US20050182793A1 (en) | 2004-02-18 | 2005-08-18 | Keenan Viktor M. | Map structure and method for producing |
US7596285B2 (en) | 2004-02-26 | 2009-09-29 | International Business Machines Corporation | Providing a portion of an electronic mail message at a reduced resolution |
US7865301B2 (en) | 2004-03-23 | 2011-01-04 | Google Inc. | Secondary map in digital mapping system |
US7599790B2 (en) | 2004-03-23 | 2009-10-06 | Google Inc. | Generating and serving tiles in a digital mapping system |
CA2820249C (en) | 2004-03-23 | 2016-07-19 | Google Inc. | A digital mapping system |
US20060026120A1 (en) | 2004-03-24 | 2006-02-02 | Update Publications Lp | Method and system for collecting, processing, and distributing residential property data |
US7269801B2 (en) | 2004-03-30 | 2007-09-11 | Autodesk, Inc. | System for managing the navigational usability of an interactive map |
JP2007535764A (en) | 2004-04-26 | 2007-12-06 | ライト90,インコーポレイテッド | Real-time data prediction |
US20050246327A1 (en) | 2004-04-30 | 2005-11-03 | Yeung Simon D | User interfaces and methods of using the same |
US20050251786A1 (en) | 2004-05-07 | 2005-11-10 | International Business Machines Corporation | System and method for dynamic software installation instructions |
US8289390B2 (en) | 2004-07-28 | 2012-10-16 | Sri International | Method and apparatus for total situational awareness and monitoring |
US7290698B2 (en) | 2004-08-25 | 2007-11-06 | Sony Corporation | Progress bar with multiple portions |
US7617232B2 (en) | 2004-09-02 | 2009-11-10 | Microsoft Corporation | Centralized terminology and glossary development |
US7933862B2 (en) | 2004-09-27 | 2011-04-26 | Microsoft Corporation | One click conditional formatting method and system for software programs |
US7788589B2 (en) | 2004-09-30 | 2010-08-31 | Microsoft Corporation | Method and system for improved electronic task flagging and management |
US7712049B2 (en) | 2004-09-30 | 2010-05-04 | Microsoft Corporation | Two-dimensional radial user interface for computer software applications |
US20060074881A1 (en) | 2004-10-02 | 2006-04-06 | Adventnet, Inc. | Structure independent searching in disparate databases |
US7284198B2 (en) | 2004-10-07 | 2007-10-16 | International Business Machines Corporation | Method and system for document draft reminder based on inactivity |
US7797197B2 (en) | 2004-11-12 | 2010-09-14 | Amazon Technologies, Inc. | Method and system for analyzing the performance of affiliate sites |
US7620628B2 (en) | 2004-12-06 | 2009-11-17 | Yahoo! Inc. | Search processing with automatic categorization of queries |
US20060129746A1 (en) | 2004-12-14 | 2006-06-15 | Ithink, Inc. | Method and graphic interface for storing, moving, sending or printing electronic data to two or more locations, in two or more formats with a single save function |
US7849395B2 (en) | 2004-12-15 | 2010-12-07 | Microsoft Corporation | Filter and sort by color |
US7451397B2 (en) | 2004-12-15 | 2008-11-11 | Microsoft Corporation | System and method for automatically completing spreadsheet formulas |
US7660823B2 (en) | 2004-12-30 | 2010-02-09 | Sas Institute Inc. | Computer-implemented system and method for visualizing OLAP and multidimensional data in a calendar format |
US9436945B2 (en) | 2005-02-01 | 2016-09-06 | Redfin Corporation | Interactive map-based search and advertising |
US7614006B2 (en) | 2005-02-11 | 2009-11-03 | International Business Machines Corporation | Methods and apparatus for implementing inline controls for transposing rows and columns of computer-based tables |
US8646080B2 (en) | 2005-09-16 | 2014-02-04 | Avg Technologies Cy Limited | Method and apparatus for removing harmful software |
US20060242630A1 (en) | 2005-03-09 | 2006-10-26 | Maxis Co., Ltd. | Process for preparing design procedure document and apparatus for the same |
US7676845B2 (en) | 2005-03-24 | 2010-03-09 | Microsoft Corporation | System and method of selectively scanning a file on a computing device for malware |
US7426654B2 (en) * | 2005-04-14 | 2008-09-16 | Verizon Business Global Llc | Method and system for providing customer controlled notifications in a managed network services system |
US7525422B2 (en) * | 2005-04-14 | 2009-04-28 | Verizon Business Global Llc | Method and system for providing alarm reporting in a managed network services environment |
US20060242040A1 (en) | 2005-04-20 | 2006-10-26 | Aim Holdings Llc | Method and system for conducting sentiment analysis for securities research |
US8639757B1 (en) | 2011-08-12 | 2014-01-28 | Sprint Communications Company L.P. | User localization using friend location information |
US8082172B2 (en) | 2005-04-26 | 2011-12-20 | The Advisory Board Company | System and method for peer-profiling individual performance |
US7958120B2 (en) | 2005-05-10 | 2011-06-07 | Netseer, Inc. | Method and apparatus for distributed community finding |
US7672968B2 (en) | 2005-05-12 | 2010-03-02 | Apple Inc. | Displaying a tooltip associated with a concurrently displayed database object |
US8024778B2 (en) | 2005-05-24 | 2011-09-20 | CRIF Corporation | System and method for defining attributes, decision rules, or both, for remote execution, claim set I |
US8825370B2 (en) | 2005-05-27 | 2014-09-02 | Yahoo! Inc. | Interactive map-based travel guide |
US8161122B2 (en) | 2005-06-03 | 2012-04-17 | Messagemind, Inc. | System and method of dynamically prioritized electronic mail graphical user interface, and measuring email productivity and collaboration trends |
US8200676B2 (en) | 2005-06-28 | 2012-06-12 | Nokia Corporation | User interface for geographic search |
US20070016363A1 (en) | 2005-07-15 | 2007-01-18 | Oracle International Corporation | Interactive map-based user interface for transportation planning |
JP3989527B2 (en) | 2005-08-04 | 2007-10-10 | 松下電器産業株式会社 | Search article estimation apparatus and method, and search article estimation apparatus server |
EP1917544A2 (en) | 2005-08-23 | 2008-05-07 | R.A. Smith & Associates, Inc. | High accuracy survey-grade gis system |
US7917841B2 (en) | 2005-08-29 | 2011-03-29 | Edgar Online, Inc. | System and method for rendering data |
JP2007079641A (en) | 2005-09-09 | 2007-03-29 | Canon Inc | Information processor and processing method, program, and storage medium |
US8095866B2 (en) | 2005-09-09 | 2012-01-10 | Microsoft Corporation | Filtering user interface for a data summary table |
US7716226B2 (en) | 2005-09-27 | 2010-05-11 | Patentratings, Llc | Method and system for probabilistically quantifying and visualizing relevance between two or more citationally or contextually related data objects |
US20070078832A1 (en) | 2005-09-30 | 2007-04-05 | Yahoo! Inc. | Method and system for using smart tags and a recommendation engine using smart tags |
US7870493B2 (en) | 2005-10-03 | 2011-01-11 | Microsoft Corporation | Distributed clipboard |
US7574428B2 (en) | 2005-10-11 | 2009-08-11 | Telmap Ltd | Geometry-based search engine for navigation systems |
US7933897B2 (en) | 2005-10-12 | 2011-04-26 | Google Inc. | Entity display priority in a distributed geographic information system |
US7487139B2 (en) | 2005-10-12 | 2009-02-03 | International Business Machines Corporation | Method and system for filtering a table |
US7627812B2 (en) | 2005-10-27 | 2009-12-01 | Microsoft Corporation | Variable formatting of cells |
US20100198858A1 (en) | 2005-11-21 | 2010-08-05 | Anti-Gang Enforcement Networking Technology, Inc. | System and Methods for Linking Multiple Events Involving Firearms and Gang Related Activities |
US7725530B2 (en) | 2005-12-12 | 2010-05-25 | Google Inc. | Proxy server collection of data for module incorporation into a container document |
US8185819B2 (en) | 2005-12-12 | 2012-05-22 | Google Inc. | Module specification for a module to be incorporated into a container document |
US7730082B2 (en) | 2005-12-12 | 2010-06-01 | Google Inc. | Remote module incorporation into a container document |
US7730109B2 (en) | 2005-12-12 | 2010-06-01 | Google, Inc. | Message catalogs for remote modules |
US9141913B2 (en) | 2005-12-16 | 2015-09-22 | Nextbio | Categorization and filtering of scientific data |
US20070150369A1 (en) | 2005-12-28 | 2007-06-28 | Zivin Michael A | Method and system for determining the optimal travel route by which customers can purchase local goods at the lowest total cost |
CN100481077C (en) | 2006-01-12 | 2009-04-22 | 国际商业机器公司 | Visual method and device for strengthening search result guide |
US7634717B2 (en) | 2006-01-23 | 2009-12-15 | Microsoft Corporation | Multiple conditional formatting |
US7818291B2 (en) | 2006-02-03 | 2010-10-19 | The General Electric Company | Data object access system and method using dedicated task object |
US7770100B2 (en) | 2006-02-27 | 2010-08-03 | Microsoft Corporation | Dynamic thresholds for conditional formats |
US20070208498A1 (en) | 2006-03-03 | 2007-09-06 | Inrix, Inc. | Displaying road traffic condition information and user controls |
US7579965B2 (en) | 2006-03-03 | 2009-08-25 | Andrew Bucholz | Vehicle data collection and processing system |
US7899611B2 (en) | 2006-03-03 | 2011-03-01 | Inrix, Inc. | Detecting anomalous road traffic conditions |
US20080052142A1 (en) | 2006-03-13 | 2008-02-28 | Bailey Maurice G T | System and method for real-time display of emergencies, resources and personnel |
DE602006002873D1 (en) | 2006-03-31 | 2008-11-06 | Research In Motion Ltd | A user interface method and apparatus for controlling the visual display of maps with selectable map elements in mobile communication devices |
US20070240062A1 (en) | 2006-04-07 | 2007-10-11 | Christena Jennifer Y | Method and System for Restricting User Operations in a Graphical User Inerface Window |
US8739278B2 (en) | 2006-04-28 | 2014-05-27 | Oracle International Corporation | Techniques for fraud monitoring and detection using application fingerprinting |
US7657626B1 (en) | 2006-09-19 | 2010-02-02 | Enquisite, Inc. | Click fraud detection |
US7468662B2 (en) | 2006-06-16 | 2008-12-23 | International Business Machines Corporation | Method for spatio-temporal event detection using composite definitions for camera systems |
US8290943B2 (en) | 2006-07-14 | 2012-10-16 | Raytheon Company | Geographical information display system and method |
WO2008022051A2 (en) | 2006-08-10 | 2008-02-21 | Loma Linda University Medical Center | Advanced emergency geographical information system |
US20130150004A1 (en) | 2006-08-11 | 2013-06-13 | Michael Rosen | Method and apparatus for reducing mobile phone usage while driving |
US20080040684A1 (en) | 2006-08-14 | 2008-02-14 | Richard Crump | Intelligent Pop-Up Window Method and Apparatus |
US20080077597A1 (en) | 2006-08-24 | 2008-03-27 | Lance Butler | Systems and methods for photograph mapping |
US20080051989A1 (en) | 2006-08-25 | 2008-02-28 | Microsoft Corporation | Filtering of data layered on mapping applications |
JP4778865B2 (en) | 2006-08-30 | 2011-09-21 | 株式会社ソニー・コンピュータエンタテインメント | Image viewer, image display method and program |
US7725547B2 (en) | 2006-09-06 | 2010-05-25 | International Business Machines Corporation | Informing a user of gestures made by others out of the user's line of sight |
US8271429B2 (en) | 2006-09-11 | 2012-09-18 | Wiredset Llc | System and method for collecting and processing data |
US7698336B2 (en) | 2006-10-26 | 2010-04-13 | Microsoft Corporation | Associating geographic-related information with objects |
US7792868B2 (en) | 2006-11-10 | 2010-09-07 | Microsoft Corporation | Data object linking and browsing tool |
US7962495B2 (en) | 2006-11-20 | 2011-06-14 | Palantir Technologies, Inc. | Creating data in a data store using a dynamic ontology |
US8069202B1 (en) | 2007-02-02 | 2011-11-29 | Resource Consortium Limited | Creating a projection of a situational network |
US7680939B2 (en) | 2006-12-20 | 2010-03-16 | Yahoo! Inc. | Graphical user interface to manipulate syndication data feeds |
US7809703B2 (en) | 2006-12-22 | 2010-10-05 | International Business Machines Corporation | Usage of development context in search operations |
US20080162616A1 (en) | 2006-12-29 | 2008-07-03 | Sap Ag | Skip relation pattern for graph structures |
US8368695B2 (en) | 2007-02-08 | 2013-02-05 | Microsoft Corporation | Transforming offline maps into interactive online maps |
US7920963B2 (en) | 2007-02-22 | 2011-04-05 | Iac Search & Media, Inc. | Map interface with a movable marker |
US8352881B2 (en) | 2007-03-08 | 2013-01-08 | International Business Machines Corporation | Method, apparatus and program storage device for providing customizable, immediate and radiating menus for accessing applications and actions |
JP5268274B2 (en) | 2007-03-30 | 2013-08-21 | キヤノン株式会社 | Search device, method, and program |
US8229458B2 (en) | 2007-04-08 | 2012-07-24 | Enhanced Geographic Llc | Systems and methods to determine the name of a location visited by a user of a wireless device |
US20080255973A1 (en) | 2007-04-10 | 2008-10-16 | Robert El Wade | Sales transaction analysis tool and associated method of use |
AU2008242910A1 (en) | 2007-04-17 | 2008-10-30 | Emd Millipore Corporation | Graphical user interface for analysis and comparison of location-specific multiparameter data sets |
US8312546B2 (en) | 2007-04-23 | 2012-11-13 | Mcafee, Inc. | Systems, apparatus, and methods for detecting malware |
US20080267107A1 (en) | 2007-04-27 | 2008-10-30 | Outland Research, Llc | Attraction wait-time inquiry apparatus, system and method |
DE102008010419A1 (en) | 2007-05-03 | 2008-11-13 | Navigon Ag | Apparatus and method for creating a text object |
US8090603B2 (en) | 2007-05-11 | 2012-01-03 | Fansnap, Inc. | System and method for selecting event tickets |
WO2009038822A2 (en) | 2007-05-25 | 2009-03-26 | The Research Foundation Of State University Of New York | Spectral clustering for multi-type relational data |
US8515207B2 (en) | 2007-05-25 | 2013-08-20 | Google Inc. | Annotations in panoramic images, and applications thereof |
US8739123B2 (en) | 2007-05-28 | 2014-05-27 | Google Inc. | Incorporating gadget functionality on webpages |
US7809785B2 (en) | 2007-05-28 | 2010-10-05 | Google Inc. | System using router in a web browser for inter-domain communication |
US20090027418A1 (en) | 2007-07-24 | 2009-01-29 | Maru Nimit H | Map-based interfaces for storing and locating information about geographical areas |
US8234298B2 (en) | 2007-07-25 | 2012-07-31 | International Business Machines Corporation | System and method for determining driving factor in a data cube |
US10762080B2 (en) | 2007-08-14 | 2020-09-01 | John Nicholas and Kristin Gross Trust | Temporal document sorter and method |
US20090055251A1 (en) | 2007-08-20 | 2009-02-26 | Weblistic, Inc., A California Corporation | Directed online advertising system and method |
US8631015B2 (en) | 2007-09-06 | 2014-01-14 | Linkedin Corporation | Detecting associates |
US20090088964A1 (en) | 2007-09-28 | 2009-04-02 | Dave Schaaf | Map scrolling method and apparatus for navigation system for selectively displaying icons |
US20090125369A1 (en) | 2007-10-26 | 2009-05-14 | Crowe Horwath Llp | System and method for analyzing and dispositioning money laundering suspicious activity alerts |
US8200618B2 (en) | 2007-11-02 | 2012-06-12 | International Business Machines Corporation | System and method for analyzing data in a report |
US20090126020A1 (en) | 2007-11-09 | 2009-05-14 | Norton Richard Elliott | Engine for rule based content filtering |
WO2009061501A1 (en) | 2007-11-09 | 2009-05-14 | Telecommunication Systems, Inc. | Points-of-interest panning on a displayed map with a persistent search on a wireless phone |
KR20090050577A (en) | 2007-11-16 | 2009-05-20 | 삼성전자주식회사 | User interface for displaying and playing multimedia contents and apparatus comprising the same and control method thereof |
US8145703B2 (en) | 2007-11-16 | 2012-03-27 | Iac Search & Media, Inc. | User interface and method in a local search system with related search results |
US20090132953A1 (en) | 2007-11-16 | 2009-05-21 | Iac Search & Media, Inc. | User interface and method in local search system with vertical search results and an interactive map |
US20090144262A1 (en) | 2007-12-04 | 2009-06-04 | Microsoft Corporation | Search query transformation using direct manipulation |
US8001482B2 (en) | 2007-12-21 | 2011-08-16 | International Business Machines Corporation | Method of displaying tab titles |
US8230333B2 (en) | 2007-12-26 | 2012-07-24 | Vistracks, Inc. | Analysis of time-based geospatial mashups using AD HOC visual queries |
US7865308B2 (en) | 2007-12-28 | 2011-01-04 | Yahoo! Inc. | User-generated activity maps |
US8010886B2 (en) | 2008-01-04 | 2011-08-30 | Microsoft Corporation | Intelligently representing files in a view |
US7805457B1 (en) | 2008-02-14 | 2010-09-28 | Securus Technologies, Inc. | System and method for identifying members of a gang or security threat group |
WO2009115921A2 (en) * | 2008-02-22 | 2009-09-24 | Ipath Technologies Private Limited | Techniques for enterprise resource mobilization |
US20090222760A1 (en) | 2008-02-29 | 2009-09-03 | Halverson Steven G | Method, System and Computer Program Product for Automating the Selection and Ordering of Column Data in a Table for a User |
US20090234720A1 (en) | 2008-03-15 | 2009-09-17 | Gridbyte | Method and System for Tracking and Coaching Service Professionals |
US9830366B2 (en) | 2008-03-22 | 2017-11-28 | Thomson Reuters Global Resources | Online analytic processing cube with time stamping |
WO2009132106A2 (en) | 2008-04-22 | 2009-10-29 | Oxford J Craig | System and method for interactive map, database, and social networking engine |
US8620641B2 (en) | 2008-05-16 | 2013-12-31 | Blackberry Limited | Intelligent elision |
US8452790B1 (en) | 2008-06-13 | 2013-05-28 | Ustringer LLC | Method and apparatus for distributing content |
US8301904B1 (en) | 2008-06-24 | 2012-10-30 | Mcafee, Inc. | System, method, and computer program product for automatically identifying potentially unwanted data as unwanted |
AU2009266403A1 (en) | 2008-07-02 | 2010-01-07 | Pacific Knowledge Systems Pty. Ltd. | Method and system for generating text |
WO2010006334A1 (en) | 2008-07-11 | 2010-01-14 | Videosurf, Inc. | Apparatus and software system for and method of performing a visual-relevance-rank subsequent search |
GB2461771A (en) | 2008-07-11 | 2010-01-20 | Icyte Pty Ltd | Annotation of electronic documents with preservation of document as originally annotated |
US8301464B1 (en) | 2008-07-18 | 2012-10-30 | Cave Consulting Group, Inc. | Method and system for producing statistical analysis of medical care information |
WO2010017229A1 (en) | 2008-08-04 | 2010-02-11 | Younoodle, Inc. | Entity performance analysis engines |
US8010545B2 (en) | 2008-08-28 | 2011-08-30 | Palo Alto Research Center Incorporated | System and method for providing a topic-directed search |
US20110078055A1 (en) | 2008-09-05 | 2011-03-31 | Claude Faribault | Methods and systems for facilitating selecting and/or purchasing of items |
US8041714B2 (en) | 2008-09-15 | 2011-10-18 | Palantir Technologies, Inc. | Filter chains with associated views for exploring large data sets |
US9348499B2 (en) | 2008-09-15 | 2016-05-24 | Palantir Technologies, Inc. | Sharing objects that rely on local resources with outside servers |
US20100070845A1 (en) | 2008-09-17 | 2010-03-18 | International Business Machines Corporation | Shared web 2.0 annotations linked to content segments of web documents |
US8214361B1 (en) | 2008-09-30 | 2012-07-03 | Google Inc. | Organizing search results in a topic hierarchy |
US8554579B2 (en) * | 2008-10-13 | 2013-10-08 | Fht, Inc. | Management, reporting and benchmarking of medication preparation |
US20100114887A1 (en) | 2008-10-17 | 2010-05-06 | Google Inc. | Textual Disambiguation Using Social Connections |
US8108933B2 (en) | 2008-10-21 | 2012-01-31 | Lookout, Inc. | System and method for attack and malware prevention |
US8411046B2 (en) | 2008-10-23 | 2013-04-02 | Microsoft Corporation | Column organization of content |
US20100131457A1 (en) | 2008-11-26 | 2010-05-27 | Microsoft Corporation | Flattening multi-dimensional data sets into de-normalized form |
US8762869B2 (en) | 2008-12-23 | 2014-06-24 | Intel Corporation | Reduced complexity user interface |
US20100191563A1 (en) | 2009-01-23 | 2010-07-29 | Doctors' Administrative Solutions, Llc | Physician Practice Optimization Tracking |
US8601401B2 (en) | 2009-01-30 | 2013-12-03 | Navico Holding As | Method, apparatus and computer program product for synchronizing cursor events |
US9177264B2 (en) | 2009-03-06 | 2015-11-03 | Chiaramail, Corp. | Managing message categories in a network |
US8447722B1 (en) | 2009-03-25 | 2013-05-21 | Mcafee, Inc. | System and method for data mining and security policy management |
IL197961A0 (en) | 2009-04-05 | 2009-12-24 | Guy Shaked | Methods for effective processing of time series |
US9767427B2 (en) | 2009-04-30 | 2017-09-19 | Hewlett Packard Enterprise Development Lp | Modeling multi-dimensional sequence data over streams |
US8719249B2 (en) | 2009-05-12 | 2014-05-06 | Microsoft Corporation | Query classification |
US8856691B2 (en) | 2009-05-29 | 2014-10-07 | Microsoft Corporation | Gesture tool |
US9268761B2 (en) | 2009-06-05 | 2016-02-23 | Microsoft Technology Licensing, Llc | In-line dynamic text with variable formatting |
US20100321399A1 (en) | 2009-06-18 | 2010-12-23 | Patrik Ellren | Maps from Sparse Geospatial Data Tiles |
KR101076887B1 (en) | 2009-06-26 | 2011-10-25 | 주식회사 하이닉스반도체 | Method of fabricating landing plug in semiconductor device |
US8635223B2 (en) | 2009-07-28 | 2014-01-21 | Fti Consulting, Inc. | System and method for providing a classification suggestion for electronically stored information |
WO2011020101A2 (en) | 2009-08-14 | 2011-02-17 | Telogis, Inc. | Real time map rendering with data clustering and expansion and overlay |
US8560548B2 (en) | 2009-08-19 | 2013-10-15 | International Business Machines Corporation | System, method, and apparatus for multidimensional exploration of content items in a content store |
US8334773B2 (en) * | 2009-08-28 | 2012-12-18 | Deal Magic, Inc. | Asset monitoring and tracking system |
JP5431235B2 (en) | 2009-08-28 | 2014-03-05 | 株式会社日立製作所 | Equipment condition monitoring method and apparatus |
US9280777B2 (en) | 2009-09-08 | 2016-03-08 | Target Brands, Inc. | Operations dashboard |
US8756489B2 (en) | 2009-09-17 | 2014-06-17 | Adobe Systems Incorporated | Method and system for dynamic assembly of form fragments |
US20110074811A1 (en) | 2009-09-25 | 2011-03-31 | Apple Inc. | Map Layout for Print Production |
US20110078173A1 (en) | 2009-09-30 | 2011-03-31 | Avaya Inc. | Social Network User Interface |
US8554699B2 (en) | 2009-10-20 | 2013-10-08 | Google Inc. | Method and system for detecting anomalies in time series data |
US8312367B2 (en) | 2009-10-30 | 2012-11-13 | Synopsys, Inc. | Technique for dynamically sizing columns in a table |
EP2499748A4 (en) | 2009-11-13 | 2017-03-01 | Zoll Medical Corporation | Community-based response system |
US11122009B2 (en) | 2009-12-01 | 2021-09-14 | Apple Inc. | Systems and methods for identifying geographic locations of social media content collected over social networks |
US20110153384A1 (en) | 2009-12-17 | 2011-06-23 | Matthew Donald Horne | Visual comps builder |
US8676597B2 (en) | 2009-12-28 | 2014-03-18 | General Electric Company | Methods and systems for mapping healthcare services analytics for volume and trends |
US8564596B2 (en) | 2010-01-12 | 2013-10-22 | Palantir Technologies, Inc. | Techniques for density mapping |
US20110167710A1 (en) | 2010-01-14 | 2011-07-14 | Boys Donald R | Trap |
US8271461B2 (en) | 2010-01-18 | 2012-09-18 | Battelle Memorial Institute | Storing and managing information artifacts collected by information analysts using a computing device |
US8683363B2 (en) | 2010-01-26 | 2014-03-25 | Apple Inc. | Device, method, and graphical user interface for managing user interface content and user interface elements |
US20110218934A1 (en) | 2010-03-03 | 2011-09-08 | Jeremy Elser | System and methods for comparing real properties for purchase and for generating heat maps to aid in identifying price anomalies of such real properties |
US8863279B2 (en) | 2010-03-08 | 2014-10-14 | Raytheon Company | System and method for malware detection |
US20110231296A1 (en) | 2010-03-16 | 2011-09-22 | UberMedia, Inc. | Systems and methods for interacting with messages, authors, and followers |
US8577911B1 (en) | 2010-03-23 | 2013-11-05 | Google Inc. | Presenting search term refinements |
US20110238553A1 (en) | 2010-03-26 | 2011-09-29 | Ashwin Raj | Electronic account-to-account funds transfer |
US8572023B2 (en) | 2010-04-14 | 2013-10-29 | Bank Of America Corporation | Data services framework workflow processing |
US8799812B2 (en) | 2010-04-29 | 2014-08-05 | Cheryl Parker | System and method for geographic based data visualization and extraction |
US8489331B2 (en) | 2010-04-29 | 2013-07-16 | Microsoft Corporation | Destination maps user interface |
US8595234B2 (en) | 2010-05-17 | 2013-11-26 | Wal-Mart Stores, Inc. | Processing data feeds |
US20110289407A1 (en) | 2010-05-18 | 2011-11-24 | Naik Devang K | Font recommendation engine |
US20110289397A1 (en) | 2010-05-19 | 2011-11-24 | Mauricio Eastmond | Displaying Table Data in a Limited Display Area |
US8723679B2 (en) | 2010-05-25 | 2014-05-13 | Public Engines, Inc. | Systems and methods for transmitting alert messages relating to events that occur within a pre-defined area |
US8756224B2 (en) | 2010-06-16 | 2014-06-17 | Rallyverse, Inc. | Methods, systems, and media for content ranking using real-time data |
US20110310005A1 (en) | 2010-06-17 | 2011-12-22 | Qualcomm Incorporated | Methods and apparatus for contactless gesture recognition |
US8489641B1 (en) | 2010-07-08 | 2013-07-16 | Google Inc. | Displaying layers of search results on a map |
US20120019559A1 (en) | 2010-07-20 | 2012-01-26 | Siler Lucas C | Methods and Apparatus for Interactive Display of Images and Measurements |
DE102010036906A1 (en) | 2010-08-06 | 2012-02-09 | Tavendo Gmbh | Configurable pie menu |
US20120036013A1 (en) | 2010-08-09 | 2012-02-09 | Brent Lee Neuhaus | System and method for determining a consumer's location code from payment transaction data |
US20120050293A1 (en) | 2010-08-25 | 2012-03-01 | Apple, Inc. | Dynamically smoothing a curve |
US8661335B2 (en) | 2010-09-20 | 2014-02-25 | Blackberry Limited | Methods and systems for identifying content elements |
US8463036B1 (en) | 2010-09-30 | 2013-06-11 | A9.Com, Inc. | Shape-based search of a collection of content |
EP2444134A1 (en) | 2010-10-19 | 2012-04-25 | Travian Games GmbH | Methods, server system and browser clients for providing a game map of a browser-based online multi-player game |
US8781169B2 (en) | 2010-11-03 | 2014-07-15 | Endeavoring, Llc | Vehicle tracking and locating system |
US8316030B2 (en) | 2010-11-05 | 2012-11-20 | Nextgen Datacom, Inc. | Method and system for document classification or search using discrete words |
JP5706137B2 (en) | 2010-11-22 | 2015-04-22 | インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation | Method and computer program for displaying a plurality of posts (groups of data) on a computer screen in real time along a plurality of axes |
WO2012071571A2 (en) | 2010-11-26 | 2012-05-31 | Agency For Science, Technology And Research | Method for creating a report from radiological images using electronic report templates |
US8839133B2 (en) | 2010-12-02 | 2014-09-16 | Microsoft Corporation | Data visualizations including interactive time line representations |
US9141405B2 (en) | 2010-12-15 | 2015-09-22 | International Business Machines Corporation | User interface construction |
US20120159399A1 (en) | 2010-12-17 | 2012-06-21 | International Business Machines Corporation | System for organizing and navigating data within a table |
US9378294B2 (en) | 2010-12-17 | 2016-06-28 | Microsoft Technology Licensing, Llc | Presenting source regions of rendered source web pages in target regions of target web pages |
US9881257B2 (en) | 2010-12-29 | 2018-01-30 | Tickr, Inc. | Multi-dimensional visualization of temporal information |
US8437731B2 (en) | 2011-01-28 | 2013-05-07 | Don Reich | Emergency call analysis system |
US8447263B2 (en) | 2011-01-28 | 2013-05-21 | Don Reich | Emergency call analysis system |
KR101950529B1 (en) | 2011-02-24 | 2019-02-20 | 렉시스넥시스, 어 디비젼 오브 리드 엘서비어 인크. | Methods for electronic document searching and graphically representing electronic document searches |
US20120246148A1 (en) * | 2011-03-22 | 2012-09-27 | Intergraph Technologies Company | Contextual Display and Scrolling of Search Results in Graphical Environment |
US9449010B2 (en) | 2011-04-02 | 2016-09-20 | Open Invention Network, Llc | System and method for managing sensitive data using intelligent mobile agents on a network |
US20120310702A1 (en) | 2011-06-03 | 2012-12-06 | Uc Group Limited | Systems and methods for monitoring compulsive behavior and for identifying early warning indicators across multiple websites |
US9104765B2 (en) | 2011-06-17 | 2015-08-11 | Robert Osann, Jr. | Automatic webpage characterization and search results annotation |
US20130006725A1 (en) | 2011-06-30 | 2013-01-03 | Accenture Global Services Limited | Tolling integration technology |
US9026944B2 (en) | 2011-07-14 | 2015-05-05 | Microsoft Technology Licensing, Llc | Managing content through actions on context based menus |
US8726379B1 (en) | 2011-07-15 | 2014-05-13 | Norse Corporation | Systems and methods for dynamic protection from electronic attacks |
US8751399B2 (en) | 2011-07-15 | 2014-06-10 | Wal-Mart Stores, Inc. | Multi-channel data driven, real-time anti-money laundering system for electronic payment cards |
US8666919B2 (en) | 2011-07-29 | 2014-03-04 | Accenture Global Services Limited | Data quality management for profiling, linking, cleansing and migrating data |
US8533204B2 (en) | 2011-09-02 | 2013-09-10 | Xerox Corporation | Text-based searching of image data |
US10031646B2 (en) | 2011-09-07 | 2018-07-24 | Mcafee, Llc | Computer system security dashboard |
US10140620B2 (en) | 2011-09-15 | 2018-11-27 | Stephan HEATH | Mobile device system and method providing combined delivery system using 3D geo-target location-based mobile commerce searching/purchases, discounts/coupons products, goods, and services, or service providers-geomapping-company/local and socially-conscious information/social networking (“PS-GM-C/LandSC/I-SN”) |
US8903355B2 (en) | 2011-09-26 | 2014-12-02 | Solacom Technologies Inc. | Answering or releasing emergency calls from a map display for an emergency services platform |
US20130097482A1 (en) | 2011-10-13 | 2013-04-18 | Microsoft Corporation | Search result entry truncation using pixel-based approximation |
US20130101159A1 (en) | 2011-10-21 | 2013-04-25 | Qualcomm Incorporated | Image and video based pedestrian traffic estimation |
US9460218B2 (en) | 2011-10-26 | 2016-10-04 | Google Inc. | Indicating location status |
US9411797B2 (en) | 2011-10-31 | 2016-08-09 | Microsoft Technology Licensing, Llc | Slicer elements for filtering tabular data |
US9053083B2 (en) | 2011-11-04 | 2015-06-09 | Microsoft Technology Licensing, Llc | Interaction between web gadgets and spreadsheets |
US8498984B1 (en) | 2011-11-21 | 2013-07-30 | Google Inc. | Categorization of search results |
US9026364B2 (en) | 2011-12-12 | 2015-05-05 | Toyota Jidosha Kabushiki Kaisha | Place affinity estimation |
US20130157234A1 (en) | 2011-12-14 | 2013-06-20 | Microsoft Corporation | Storyline visualization |
US20130166550A1 (en) | 2011-12-21 | 2013-06-27 | Sap Ag | Integration of Tags and Object Data |
US9189556B2 (en) | 2012-01-06 | 2015-11-17 | Google Inc. | System and method for displaying information local to a selected area |
WO2013102892A1 (en) | 2012-01-06 | 2013-07-11 | Technologies Of Voice Interface Ltd | A system and method for generating personalized sensor-based activation of software |
US9116994B2 (en) | 2012-01-09 | 2015-08-25 | Brightedge Technologies, Inc. | Search engine optimization for category specific search results |
AU2013222093A1 (en) | 2012-02-24 | 2014-09-11 | Mccormick & Company, Incorporated | System and method for providing flavor advisement and enhancement |
WO2013130633A1 (en) | 2012-02-29 | 2013-09-06 | Google Inc. | Interactive query completion templates |
US8787939B2 (en) | 2012-03-27 | 2014-07-22 | Facebook, Inc. | Dynamic geographic beacons for geographic-positioning-capable devices |
US20130263019A1 (en) | 2012-03-30 | 2013-10-03 | Maria G. Castellanos | Analyzing social media |
US8738665B2 (en) | 2012-04-02 | 2014-05-27 | Apple Inc. | Smart progress indicator |
US8983936B2 (en) | 2012-04-04 | 2015-03-17 | Microsoft Corporation | Incremental visualization for structured data in an enterprise-level data store |
US9071653B2 (en) | 2012-04-05 | 2015-06-30 | Verizon Patent And Licensing Inc. | Reducing cellular network traffic |
US8792677B2 (en) | 2012-04-19 | 2014-07-29 | Intelligence Based Integrated Security Systems, Inc. | Large venue security method |
US9298856B2 (en) | 2012-04-23 | 2016-03-29 | Sap Se | Interactive data exploration and visualization tool |
US9043710B2 (en) | 2012-04-26 | 2015-05-26 | Sap Se | Switch control in report generation |
US8742934B1 (en) | 2012-04-29 | 2014-06-03 | Intel-Based Solutions, LLC | System and method for facilitating the execution of law enforcement duties and enhancing anti-terrorism and counter-terrorism capabilities |
US10304036B2 (en) | 2012-05-07 | 2019-05-28 | Nasdaq, Inc. | Social media profiling for one or more authors using one or more social media platforms |
US20140032506A1 (en) | 2012-06-12 | 2014-01-30 | Quality Attributes Software, Inc. | System and methods for real-time detection, correction, and transformation of time series data |
US8966441B2 (en) | 2012-07-12 | 2015-02-24 | Oracle International Corporation | Dynamic scripts to extend static applications |
US8836788B2 (en) | 2012-08-06 | 2014-09-16 | Cloudparc, Inc. | Controlling use of parking spaces and restricted locations using multiple cameras |
US8554875B1 (en) | 2012-08-13 | 2013-10-08 | Ribbon Labs, Inc. | Communicating future locations in a social network |
US10311062B2 (en) | 2012-08-21 | 2019-06-04 | Microsoft Technology Licensing, Llc | Filtering structured data using inexact, culture-dependent terms |
US8676857B1 (en) | 2012-08-23 | 2014-03-18 | International Business Machines Corporation | Context-based search for a data store related to a graph node |
US20140068487A1 (en) | 2012-09-05 | 2014-03-06 | Roche Diagnostics Operations, Inc. | Computer Implemented Methods For Visualizing Correlations Between Blood Glucose Data And Events And Apparatuses Thereof |
US20140095273A1 (en) | 2012-09-28 | 2014-04-03 | Catalina Marketing Corporation | Basket aggregator and locator |
US20140095509A1 (en) | 2012-10-02 | 2014-04-03 | Banjo, Inc. | Method of tagging content lacking geotags with a location |
GB2578839B (en) | 2012-10-08 | 2020-08-19 | Fisher Rosemount Systems Inc | Dynamically reusable classes |
US9741138B2 (en) * | 2012-10-10 | 2017-08-22 | International Business Machines Corporation | Node cluster relationships in a graph database |
US9104786B2 (en) | 2012-10-12 | 2015-08-11 | International Business Machines Corporation | Iterative refinement of cohorts using visual exploration and data analytics |
US20140108068A1 (en) | 2012-10-17 | 2014-04-17 | Jonathan A. Williams | System and Method for Scheduling Tee Time |
US9501799B2 (en) | 2012-11-08 | 2016-11-22 | Hartford Fire Insurance Company | System and method for determination of insurance classification of entities |
US9146969B2 (en) | 2012-11-26 | 2015-09-29 | The Boeing Company | System and method of reduction of irrelevant information during search |
US20140157172A1 (en) | 2012-11-30 | 2014-06-05 | Drillmap | Geographic layout of petroleum drilling data and methods for processing data |
US20140156527A1 (en) | 2012-11-30 | 2014-06-05 | Bank Of America Corporation | Pre-payment authorization categorization |
US10282088B2 (en) * | 2012-12-06 | 2019-05-07 | Samsung Electronics Co., Ltd. | Configuration of application execution spaces and sub-spaces for sharing data on a mobile tough screen device |
US9497289B2 (en) | 2012-12-07 | 2016-11-15 | Genesys Telecommunications Laboratories, Inc. | System and method for social message classification based on influence |
US9294576B2 (en) | 2013-01-02 | 2016-03-22 | Microsoft Technology Licensing, Llc | Social media impact assessment |
US20140195515A1 (en) | 2013-01-10 | 2014-07-10 | I3 Analytics | Methods and systems for querying and displaying data using interactive three-dimensional representations |
US8917274B2 (en) | 2013-03-15 | 2014-12-23 | Palantir Technologies Inc. | Event matrix based on integrated data |
US9501202B2 (en) | 2013-03-15 | 2016-11-22 | Palantir Technologies, Inc. | Computer graphical user interface with genomic workflow |
US9740369B2 (en) | 2013-03-15 | 2017-08-22 | Palantir Technologies Inc. | Systems and methods for providing a tagging interface for external content |
US8937619B2 (en) | 2013-03-15 | 2015-01-20 | Palantir Technologies Inc. | Generating an object time series from data objects |
US8868486B2 (en) | 2013-03-15 | 2014-10-21 | Palantir Technologies Inc. | Time-sensitive cube |
US8909656B2 (en) * | 2013-03-15 | 2014-12-09 | Palantir Technologies Inc. | Filter chains with associated multipath views for exploring large data sets |
GB2542517B (en) | 2013-05-07 | 2018-01-24 | Palantir Technologies Inc | Interactive Geospatial map |
US8799799B1 (en) | 2013-05-07 | 2014-08-05 | Palantir Technologies Inc. | Interactive geospatial map |
US20150019394A1 (en) | 2013-07-11 | 2015-01-15 | Mastercard International Incorporated | Merchant information correction through transaction history or detail |
US8620790B2 (en) | 2013-07-11 | 2013-12-31 | Scvngr | Systems and methods for dynamic transaction-payment routing |
US9223773B2 (en) | 2013-08-08 | 2015-12-29 | Palatir Technologies Inc. | Template system for custom document generation |
GB2518745A (en) | 2013-08-08 | 2015-04-01 | Palantir Technologies Inc | Template system for custom document generation |
US9565152B2 (en) | 2013-08-08 | 2017-02-07 | Palantir Technologies Inc. | Cable reader labeling |
US9477372B2 (en) | 2013-08-08 | 2016-10-25 | Palantir Technologies Inc. | Cable reader snippets and postboard |
US9335897B2 (en) | 2013-08-08 | 2016-05-10 | Palantir Technologies Inc. | Long click display of a context menu |
US8713467B1 (en) * | 2013-08-09 | 2014-04-29 | Palantir Technologies, Inc. | Context-sensitive views |
US8689108B1 (en) | 2013-09-24 | 2014-04-01 | Palantir Technologies, Inc. | Presentation and analysis of user interaction data |
US9785317B2 (en) | 2013-09-24 | 2017-10-10 | Palantir Technologies Inc. | Presentation and analysis of user interaction data |
US8938686B1 (en) | 2013-10-03 | 2015-01-20 | Palantir Technologies Inc. | Systems and methods for analyzing performance of an entity |
US8812960B1 (en) | 2013-10-07 | 2014-08-19 | Palantir Technologies Inc. | Cohort-based presentation of user interaction data |
US8924872B1 (en) | 2013-10-18 | 2014-12-30 | Palantir Technologies Inc. | Overview user interface of emergency call data of a law enforcement agency |
US9116975B2 (en) | 2013-10-18 | 2015-08-25 | Palantir Technologies Inc. | Systems and user interfaces for dynamic and interactive simultaneous querying of multiple data stores |
US8832594B1 (en) | 2013-11-04 | 2014-09-09 | Palantir Technologies Inc. | Space-optimized display of multi-column tables with selective text truncation based on a combined text width |
US9021384B1 (en) * | 2013-11-04 | 2015-04-28 | Palantir Technologies Inc. | Interactive vehicle information map |
US8868537B1 (en) | 2013-11-11 | 2014-10-21 | Palantir Technologies, Inc. | Simple web search |
US9235638B2 (en) | 2013-11-12 | 2016-01-12 | International Business Machines Corporation | Document retrieval using internal dictionary-hierarchies to adjust per-subject match results |
EP2884440A1 (en) | 2013-12-16 | 2015-06-17 | Palantir Technologies, Inc. | Methods and systems for analyzing entity performance |
US9727622B2 (en) | 2013-12-16 | 2017-08-08 | Palantir Technologies, Inc. | Methods and systems for analyzing entity performance |
US9043696B1 (en) | 2014-01-03 | 2015-05-26 | Palantir Technologies Inc. | Systems and methods for visual definition of data associations |
US9009827B1 (en) | 2014-02-20 | 2015-04-14 | Palantir Technologies Inc. | Security sharing system |
US9483162B2 (en) * | 2014-02-20 | 2016-11-01 | Palantir Technologies Inc. | Relationship visualizations |
US9009171B1 (en) | 2014-05-02 | 2015-04-14 | Palantir Technologies Inc. | Systems and methods for active column filtering |
US9021260B1 (en) | 2014-07-03 | 2015-04-28 | Palantir Technologies Inc. | Malware data item analysis |
US9785773B2 (en) * | 2014-07-03 | 2017-10-10 | Palantir Technologies Inc. | Malware data item analysis |
US9454281B2 (en) * | 2014-09-03 | 2016-09-27 | Palantir Technologies Inc. | System for providing dynamic linked panels in user interface |
US10402061B2 (en) * | 2014-09-28 | 2019-09-03 | Microsoft Technology Licensing, Llc | Productivity tools for content authoring |
US9767172B2 (en) * | 2014-10-03 | 2017-09-19 | Palantir Technologies Inc. | Data aggregation and analysis system |
US9785328B2 (en) * | 2014-10-06 | 2017-10-10 | Palantir Technologies Inc. | Presentation of multivariate data on a graphical user interface of a computing system |
US9043894B1 (en) | 2014-11-06 | 2015-05-26 | Palantir Technologies Inc. | Malicious software detection in a computing system |
US9348920B1 (en) * | 2014-12-22 | 2016-05-24 | Palantir Technologies Inc. | Concept indexing among database of documents using machine learning techniques |
US9335911B1 (en) * | 2014-12-29 | 2016-05-10 | Palantir Technologies Inc. | Interactive user interface for dynamic data analysis exploration and query processing |
US10387834B2 (en) * | 2015-01-21 | 2019-08-20 | Palantir Technologies Inc. | Systems and methods for accessing and storing snapshots of a remote application in a document |
US9727560B2 (en) * | 2015-02-25 | 2017-08-08 | Palantir Technologies Inc. | Systems and methods for organizing and identifying documents via hierarchies and dimensions of tags |
US10547575B2 (en) * | 2015-11-10 | 2020-01-28 | Wrinkl, Inc. | Apparatus and method for control of access to communication channels |
US9454785B1 (en) * | 2015-07-30 | 2016-09-27 | Palantir Technologies Inc. | Systems and user interfaces for holistic, data-driven investigation of bad actor behavior based on clustering and scoring of related data |
US9996595B2 (en) * | 2015-08-03 | 2018-06-12 | Palantir Technologies, Inc. | Providing full data provenance visualization for versioned datasets |
US9485265B1 (en) * | 2015-08-28 | 2016-11-01 | Palantir Technologies Inc. | Malicious activity detection system capable of efficiently processing data accessed from databases and generating alerts for display in interactive user interfaces |
EP3477906B1 (en) * | 2017-10-26 | 2021-03-31 | Accenture Global Solutions Limited | Systems and methods for identifying and mitigating outlier network activity |
-
2013
- 2013-12-03 US US14/095,798 patent/US8713467B1/en active Active
-
2014
- 2014-04-01 US US14/242,559 patent/US9557882B2/en active Active
- 2014-08-07 EP EP14180281.9A patent/EP2838039A3/en not_active Ceased
- 2014-08-07 CA CA2858586A patent/CA2858586A1/en not_active Abandoned
- 2014-08-08 AU AU2014210655A patent/AU2014210655B2/en not_active Ceased
-
2017
- 2017-01-04 US US15/398,092 patent/US9921734B2/en active Active
-
2018
- 2018-01-24 US US15/878,914 patent/US10545655B2/en active Active
Also Published As
Publication number | Publication date |
---|---|
US20170115865A1 (en) | 2017-04-27 |
AU2014210655B2 (en) | 2018-07-26 |
US9557882B2 (en) | 2017-01-31 |
US20150046870A1 (en) | 2015-02-12 |
US9921734B2 (en) | 2018-03-20 |
US20180150213A1 (en) | 2018-05-31 |
AU2014210655A1 (en) | 2015-02-26 |
US10545655B2 (en) | 2020-01-28 |
US8713467B1 (en) | 2014-04-29 |
EP2838039A3 (en) | 2015-02-25 |
EP2838039A2 (en) | 2015-02-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10545655B2 (en) | Context-sensitive views | |
US10824632B2 (en) | System and method for generating event visualizations | |
US10996823B2 (en) | Interactive geospatial map | |
US11402985B2 (en) | Systems and user interfaces for dynamic and interactive access of, investigation of, and analysis of data objects stored in one or more databases | |
US10783686B2 (en) | Interactive data object map | |
US20230044699A1 (en) | System for providing dynamic linked panels in user interface | |
US10552998B2 (en) | System and method of generating data points from one or more data stores of data items for chart creation and manipulation | |
EP3133510A1 (en) | Interactive geospatial map | |
US20220028136A1 (en) | Ontology-backed automatic chart creation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FZDE | Dead |
Effective date: 20200831 |