US20130091471A1 - Visual search and three-dimensional results - Google Patents
Visual search and three-dimensional results Download PDFInfo
- Publication number
- US20130091471A1 US20130091471A1 US13/690,472 US201213690472A US2013091471A1 US 20130091471 A1 US20130091471 A1 US 20130091471A1 US 201213690472 A US201213690472 A US 201213690472A US 2013091471 A1 US2013091471 A1 US 2013091471A1
- Authority
- US
- United States
- Prior art keywords
- objects
- dimensional interface
- visual
- visual search
- computer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
- G06F16/2458—Special types of queries, e.g. statistical queries, fuzzy queries or distributed queries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/54—Browsing; Visualisation therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/26—Visual data mining; Browsing structured data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/28—Databases characterised by their database models, e.g. relational or object models
- G06F16/289—Object oriented databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/903—Querying
- G06F16/90335—Query processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/951—Indexing; Web crawling techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/957—Browsing optimisation, e.g. caching or content distillation
-
- G06F17/30979—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
Definitions
- search engines are configured to provide results that include one or more terms of a search query.
- Conventional search engines may use indices storing references to electronic documents and the terms included in the electronic documents to generate the results.
- the search engine includes the references to the electronic documents identified in an index having similar terms in the results.
- Some conventional search engines may provide search suggestions to a user when one or more terms are entered in a conventional search interface for the conventional search engine.
- the suggested terms may be selected and used to locate electronic documents that include the entered terms and the selected suggested terms.
- the terms entered in the search interface or the suggested terms may not correspond to an object that is sought by a user.
- the conventional search interface fails to provide adequate assistance in formulating terms for a set of objects that is sought by the user.
- a seeker hopes to complete a task upon locating the set of objects.
- the seeker may be confronted with difficulty in formulating a search query that causes the search engine to provide the set of objects that is sought.
- the seeker may be searching for any of the following: best American-made hybrid cars on the market under $25,000; child-friendly movies playing today in my area; or netbooks under $400.
- the seeker may have difficulty locating the set of objects may resort to trial and error to formulate terms for the set of objects. Because the conventional search engine only performs term match using each term in the user search query, the actual content results often doesn't exactly correspond to the set of objects that is sought by the user.
- Embodiments of the invention overcoming these and other problems in the art relate in one regard to a computer system, graphical user interface, and computer-implemented method to visually search and explore a set of objects.
- the computer system renders images associated with a set of objects and transitions between a three-dimensional interface and two-dimensional interface when narrowing the set of objects.
- the computer system allows a user to visually search through and explore a set of objects related to a user's intent.
- the computer system includes a database and a server.
- the database is configured to store objects, attributes for each object, and images associated with each object.
- the server is connected to the database.
- the server retrieves a set of images and attributes in response to a visual search request.
- the server or client may generate a graphical user interface to organize the images and attributes.
- the graphical user interface may dynamically transition between a three-dimensional representation of the set of images and a two-dimensional representation of the set of images based on the number of images remaining in the set of images.
- FIG. 1 is a network diagram that illustrates an exemplary computing system in accordance with embodiments of the invention
- FIG. 2 is a graphical user interface that illustrates a search results page having a visual search link in accordance with embodiments of the invention
- FIG. 3 is a graphical user interface that illustrates images for a set of objects in a three-dimensional representation in accordance with embodiments of the invention
- FIG. 4 is a graphical user interface that illustrates metadata surfaced for each object in a two-dimensional representation in accordance with embodiments of the invention
- FIG. 5 is a graphical user interface that illustrates a set of objects that match the criteria provided by category or refine controls in accordance with embodiments of the invention
- FIG. 6 is a graphical user interface that illustrates a result set generated in response to click action in accordance with embodiments of the invention.
- FIG. 7 is a logic diagram that illustrates a method to visually search through and explore a set of objects in accordance with embodiments of the invention.
- Embodiments of the invention provide a computer system that allows a user to visually search through and explore a set of objects included in a graphical user interface that contains the results to a user query or browsing request.
- the objects in the set of objects may correspond to, but is not limited to, shopping products, automobile products, animals, famous people, entertainment media, travel, or sports.
- the graphical user interface implements visual search and exploration by allowing the user to express ideas and search terms without having to formulate words for them.
- the graphical user interface may render a visual representation for each object and important attributes for the set of objects.
- the important attributes include attributes that users frequently use to evaluate the set of objects. Users can visually interact with the important attributes rather than formulate textual queries to narrow the set of objects.
- the graphical user interface may include tangentially related categories to provide a user with an opportunity to serendipitously discover objects or relationships that are unknown to the user.
- each object in the graphical user interface may be linked with a query that may be previewed in a search box when a pointer hovers over the object. In turn, the query may be executed by clicking on the object.
- the graphical user interface may dynamically adapt based on the number of items currently available in the set of objects. Accordingly, the computer system may be used to narrow a set of objects, to identify new items in a given field, e.g., new car models, new books, new movies in theaters—especially when the name of the objects are unknown, or to surface interesting and popular facts associated within a selected category.
- the computer system may include hardware, software, or a combination of hardware and software.
- the hardware includes processors and memories configured to execute instructions stored in the memories.
- the memories include computer-readable media that store a computer-program product having computer-useable instructions for a computer-implemented method.
- Computer-readable media include both volatile and nonvolatile media, removable and nonremovable media, and media readable by a database, a switch, and various other network devices. Network switches, routers, and related components are conventional in nature, as are means of communicating with the same.
- computer-readable media comprise computer-storage media and communications media.
- Computer-storage media, or machine-readable media include media implemented in any method or technology for storing information.
- Computer-storage media include, but are not limited to, random access memory (RAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), flash memory or other memory technology, compact-disc read only memory (CD-ROM), digital versatile discs (DVD), holographic media or other optical disc storage, magnetic cassettes, magnetic tape, magnetic disk storage, and other magnetic storage devices.
- RAM random access memory
- ROM read only memory
- EEPROM electrically erasable programmable read only memory
- flash memory or other memory technology
- CD-ROM compact-disc read only memory
- DVD digital versatile discs
- holographic media or other optical disc storage magnetic cassettes, magnetic tape, magnetic disk storage, and other magnetic storage devices.
- FIG. 1 is a network diagram that illustrates an exemplary operating environment 100 , in accordance with embodiments of the invention.
- the operating environment 100 shown in FIG. 1 is merely exemplary and is not intended to suggest any limitation as to scope or functionality. Embodiments of the invention are operable with numerous other configurations.
- the operating environment 100 includes a network 110 , a visual search engine 120 , client devices 130 , web pages 140 , an objects database 150 , and index servers 160 .
- the network 110 is configured to facilitate communication between the client devices 130 and the visual search engine 120 .
- the network 110 also facilitates communication between the objects database 150 , the visual search engine 120 , and the index servers 160 .
- the network 110 may be a communication network, such as a wireless network, local area network, wired network, or the Internet.
- the client devices 130 interact with the visual search engine 120 utilizing the network 110 .
- the visual search engine 120 provides web pages, images, videos, or other electronic documents that contain terms provided or selected by the user.
- the visual search engine 120 is communicatively connected via network 110 to the client devices 130 , the objects database 150 , and the index servers 160 .
- the visual search engine 120 is a server that generates visual representations for display on the client devices 130 .
- the visual search engine 120 receives, over network 110 , selections of terms or selections of links from client devices 130 that provide interfaces that receive interactions from users.
- the visual search engine 120 may generate a three-dimensional or two-dimensional representation to display objects that correspond to terms or links selected by the user.
- the visual representation may include, but is not limited to, timelines, maps, grids or tables that organize the objects on the display.
- the visual search engine 120 is connected to a search engine to retrieve search results in response to a query formulated and issued by the visual explore engine.
- the visual search engine 120 traverses the objects database 150 to identify objects that correspond to the selections received from the client devices.
- the objects database 150 transmits a set of objects that satisfies the selections to the visual search engine 120 .
- the set of objects are associated with object queries, web pages 140 , images, videos, or other electronic documents.
- the visual search engine 120 may also traverse the index servers 160 to locate web pages 140 that satisfy the object queries or user queries.
- the index servers 160 transmit a set of uniform resource locators (URLs) that point to web pages, images, videos or other electronic documents that satisfy the object queries or user queries.
- the visual search engine 120 formats the URLs and transmits the URLs to the client devices 130 .
- the client devices 130 are utilized by a user to generate search terms, to hover over objects, or to select links or objects, and to receive results or web pages 140 that are relevant to the search terms, the selected links, or the selected objects.
- the client devices 130 include, without limitation, personal digital assistants, smart phones, laptops, personal computers, gaming systems, set-top boxes, or any other suitable client computing device.
- the client devices 130 include user and system information storage to store user and system information on the client devices 130 .
- the user information may include search histories, cookies, and passwords.
- the system information may include internet protocol addresses, cached Web pages, and system utilization.
- the client devices 130 communicate with the visual search engine 120 to receive the results or web pages 140 that are relevant to the search terms, the selected links, or the selected objects.
- the client devices 130 execute a Silverlight browser application.
- the Silverlight browser application may provide an interactive experience when formulating search terms for an item and receiving results that response to a user request or the formulated search terms.
- the client device 130 receives objects or web pages 140 selected by the visual search engine 120
- the Silverlight browser application only renders a subset of the items that satisfy the user request.
- the Silverlight browser application may randomly populate the results with some of the additional objects by illustrating at least one object of the additional objects in each row of the display real estate of the client device 130 as the visual search engine transmits the remaining additional objects to the client device 130 .
- the web pages 140 provide details about items that interest the user.
- the web pages 140 are indexed by the index servers 160 .
- Each web page 140 may include terms or metadata.
- the terms or metadata is used by the index servers to store the web page in an appropriate location. Additionally, the web pages 140 are associated with URLs that are also stored by the index servers 160 .
- the objects database 150 stores attributes and images for each object.
- the attributes include titles, image size, image dimensions, and other metadata for the object.
- the visual search engine 120 may request one or more objects from the objects database 150 .
- the objects database 150 transmits attributes and images to the visual search engine 120 .
- the index servers 160 store web pages 140 , terms associated with each web page 140 , and uniform resource locators (URLs) corresponding to the web pages 140 .
- the visual search engine 120 may request one or more web pages 140 from the index servers 160 .
- the index servers 160 transmit the web pages to the visual search engine 120 .
- the operating environment 100 is configured with a visual search engine 120 that provides results that include web pages 140 and objects to the client devices 130 .
- the visual search engine 120 traverses the object database 150 and traverses the index servers 160 to obtain results that satisfy the requests received from the users.
- the client devices 130 render the results for display to the users.
- the visual search engine generates a graphical user interface that includes results that match terms provided by or selections selected by the user.
- the results may include URLs that point to web pages.
- the graphical user interface contains a link that reformats the results into a visual representation. The link to the visual representation may be highlighted on the graphical user interface.
- FIG. 2 is a graphical user interface 200 that illustrates a search results page having a visual search link in accordance with embodiments of the invention.
- the graphical user interface 200 includes a search box 210 and a visual search link 230 .
- the search box 210 in the graphical user interface 200 allows the user to formulate their own search terms for a search query.
- the search box 210 receives search terms provided by the user.
- the search terms are utilized by the visual search engine to obtain search results that match the terms entered in the search box 210 .
- the visual search link 230 is part of the graphical user interface 200 .
- the visual search link 230 allows the user to explore results of a search query visually. Additionally, the visual search link 230 may also allow the user to visually formulate a query without having to enter their own search terms.
- the visual representation of the search results is a three-dimensional representation.
- the visual search engine may recognize that the user hopes to complete a task associated with purchasing a used or new car.
- the results that a search engine returns may not aid the user in completing the task associated with purchasing a used or new car.
- the visual search engine ignores the results from the search engine and provides a visual display having visual representations of cars that are currently for sale in the user's location. The user may interact with the visual representation and features of the graphical user interface 200 to locate the “cars” that the user seeks.
- the visual search engine may return results that consist of web pages 140 that include the term car.
- the visual search engine may interact with the visual representation and features of the graphical user interface 200 to locate the “cars” that the user seeks.
- the visual search engine may provide a browse entry point that does not require entry of a query terms.
- the browse entry page may include a collection of objects arranged by category. Each of the objects may be represented by an image and may correspond to a visual query.
- the visual representation generated by the visual explore engine, in response to a user request, to visually explore the search results is a three-dimensional representation of the search results.
- a graphical user interface displays the three-dimensional representation and controls that are used to interact with the three-dimensional representation. In turn, the user can narrow the objects included in the three-dimensional representation.
- FIG. 3 is a graphical user interface 300 that illustrates images for a set of objects 350 a in a three-dimensional representation 350 in accordance with embodiments of the invention.
- the visual search engine generates a visual representation in graphical user interface 300 when the user clicks on the visual search link in a graphical user interface that includes URL results for search terms provided by the user.
- the graphical user interface 300 replaces the URL results with the visual representation (three-dimensional or two-dimensional) based on the number of objects 350 a and the dimensions of the objects 350 a .
- the visual search engine animates the objects as the objects are rendered in the graphical user interface 300 .
- the visual search engine When the visual search engine generates the three-dimensional representation 350 , the objects 350 a move from the background of the graphical user interface 300 and arrange themselves into an overlapping three-dimensional array of objects. In turn, the visual search engine animates controls on the left and right of the graphical user interface 300 by floating the controls onto the graphical user interface 300 from the left and right. The user can move a pointer to interact with the controls that that are on the left side and the right side of the graphical user interface 300 to manipulate the objects 350 a . In some embodiments, the visual search engine generates the visual representation in the graphical user interface 300 when the user clicks on a visual search link without requiring the user to provide search terms.
- the graphical user interface 300 includes category control 310 , count control 320 , refine controls 330 , sort control 340 , and scroll control 360 that are used to interact with the objects 350 a included in the three-dimensional representation 350 .
- the graphical user interface 300 allows the users to view all objects that match the search terms provided by the users.
- the graphical user interface 300 loads less then all of the objects but the scroll control 360 provides an infinite scroll feature that allows the users to view additional objects in the visual representation that are currently outside of the viewable region of the three-dimensional representation 350 .
- the graphical user interface 300 may load the data for all objects, but stream the images as the specific objects come into view as the scroll control 360 reveals the additional objects.
- the graphical user interface may only load a subset of the data and the corresponding images to allow the user to quickly interact with some of the objects. The additional data and images are loaded as requested via the scroll control 360 .
- the category control 310 identifies the categories associated with each object in the three-dimensional representation 350 .
- the number objects 350 a in the three-dimensional representation 350 may be reduced when the user selects the category control 310 .
- the three-dimensional representation 350 animates the objects 350 a as the visual search engine reduces the number of objects 350 a in the three-dimensional representation 350 .
- the objects 350 a that are not within the selected category drop or fade into a background of the graphical user interface 300 .
- Additional objects 350 a that satisfy the selected category replace the removed objects 350 a in the three-dimensional representation 350 . All the objects 350 a that are in the updated three-dimensional representation 350 satisfy a category selected in the category control 310 .
- a user may select sport utility vehicle (SUV) as a category when searching for cars.
- SUV sport utility vehicle
- the visual search engine updates the objects 350 a in the three-dimensional representation 350 by removing objects 350 a that are not categorized as SUVs.
- the objects 350 a that do not satisfy the user selection are animated by the visual search engine to fall out of the three-dimensional representation 350 and additional objects that are outside the viewable region of the three-dimensional representation 350 are animated to replace the removed objects 350 a.
- the visual search engine may update the category control 310 with a new link based on a previous search performed by the user.
- the visual search engine creates a link that is associated with a visual representation of objects having features similar to the specific item or branded item.
- the visual search engine When the user selects the link, the visual search engine generates an interface having objects with similar features.
- the visual search engine may include links in the category control 310 to information that is tangentially related to the previous search.
- the category control 310 may update to show links to James Bond movies currently playing in movie theaters (because James Bond drives Aston Martins), and the country England (because Aston Martins are produced in England).
- the count control 320 displays the total number of items that are in the visual representation and the number of items that are currently displayed in the viewable regions of the graphical user interface 300 .
- the count control 320 updates as the user reduces or increases the objects that are within the visual representation (e.g. three-dimensional representation 350 ).
- the count control 320 updates when the user initiates a search with new or additional search terms.
- the count control 320 updates when the user interacts with the category control 310 or refine controls 330 to manipulate the objects 350 a in the graphical user interface 300 .
- the refine controls 330 represent attributes that are shared by some of the objects 350 a that are within the visual representation.
- the number objects 350 a in the three-dimensional representation 350 may be reduced when the user selects a value for the attributes represented by the refine controls 330 .
- the three-dimensional representation 350 animates the objects 350 a as the visual search engine reduces the number of objects 350 a in the three-dimensional representation 350 .
- the objects 350 a that fail to satisfy the value selected for the attributes represented by the refine controls 330 drop or fade into a background of the graphical user interface 300 . Additional objects 350 a that satisfy the value selected for the attributes represented by the refine controls 330 replace the removed objects 350 a . All the objects 350 a that are in the updated visual representation (e.g.
- the updated three-dimensional representation 350 satisfy the values selected for the attributes represented by the refine controls 330 .
- the attributes in the refine controls 350 are updated to provide only valid filtering options. For instance, when a user selects “Sport Utility Vehicle” as the vehicle type filter, the brand filter may automatically to remove manufacturers or brands that do not offer a “Sport Utility Vehicle.”
- the sort control 340 allows the user to sort the objects 350 a .
- the sort control 340 reorganizes the objects 350 a in the three-dimensional representation 350 .
- the sort control 330 represents attributes that are shared by all the objects 350 a in the visual representation.
- the objects 350 a are sorted alphabetically by default.
- the visual search engine can reorganize the objects 350 a in the visual representation. For instance, when the user is searching for cars, the sort control 340 may allow the user to sort the objects 350 a by brand, safety rating, or expert rating.
- the visual search engine obtains the brand, safety rating, and expert rating from the objects database, which stores attributes for brand, safety rating, and expert rating for the objects.
- the visual search engine reorders the objects in the visual representation.
- the attributes that are included in sort control are selected based on search frequency in a search log.
- the scroll control 360 allows the user to view additional objects 350 a that are within the visual representation.
- the scroll control provides infinite scroll, which allows the graphical user interface 300 to continuously scroll through the objects 350 a in the visual representation. The infinite scroll continuously loops through all objects 350 a within the visual representation.
- the visual search engine transitions between a three-dimensional representation 350 of the objects 350 a and a two-dimensional representation.
- the visual search engine generates the three dimensional representation 350 when the number of objects that are selected for display in the graphical user interface is large.
- the visual search engine may render the objects in a two-dimensional representation.
- the remaining objects are reoriented by the visual search engine in rows that are viewable without scrolling.
- the visual search engine may render a boundary around each of the remaining objects 350 a.
- the visual search engine may set a display default that requires generating a graphical user interface with detailed information for each object.
- the graphical user interface must also display as many of the objects that satisfy the user selection in the visual representation.
- the three-dimensional visual representation 350 a of the objects may be generated by the visual search engine when the number of objects is large and showing detailed information would obscure the graphical user interface.
- the two-dimensional visual representation may be generated by the visual search engine when the number of objects is small and showing detailed information would not obscure the graphical user interface. Additionally, because the available display real estate increases, the visual search engine displays more information for each object.
- images associated with each object are sized to 100 pixel by 100 pixels and the two-dimensional representation includes an array having a minimum of five rows and five columns of images.
- a two-dimensional representation of the objects is generated by the visual explore engine.
- the two-dimensional representation allows the user to view additional information about each object on the graphical user interface.
- the visual search engine uses the available display real estate on the graphical user interface to maximize information that is rendered for display to the user.
- FIG. 4 is a graphical user interface 400 that illustrates metadata surfaced for each object in a two-dimensional representation 440 in accordance with embodiments of the invention.
- the visual search engine generates the two-dimensional representation 440 in graphical user interface 400 based on the number of objects that are selected for rendering.
- the visual search engine may generate the two-dimensional representation 440 in graphical user interface 400 when the number of objects that are selected for display is viewable without scrolling.
- the visual search engine may reformat the images associated with the objects to fit in the graphical user interface 400 without scrolling.
- the visual search engine may generate the two-dimensional representation 440 in graphical user interface 400 when the number of objects allows for comparison of the objects based on the available metadata that is viewable on the graphical user interface 400 .
- the graphical user interface 400 includes search box 410 , pointer 420 , metadata displays 430 , the two-dimensional representation 440 , and object labels 450 .
- the search box 410 allows the user to enter search terms that should be sent to the visual explore engine.
- the search box 410 includes a search icon that causes the visual search engine to initiate a search with the search terms contained within the search box 410 .
- the search box 410 also automatically displays search queries associated with each object in the two-dimensional representation 440 when a user hovers over an object with the pointer 420 . The user may initiate a search using the search queries that are automatically displayed in the search box 410 in response to the hover by clicking on the search icon included in the search box 410 .
- the user may initiate a search by using the search queries associated with each object by clicking on the object with the pointer 420 .
- the visual search engine searches the index server and objects database using the search queries associated with the object.
- the pointer 420 is a mouse that allows the user to hover over objects, click on objects, or drag and drop objects.
- the graphical user interface 400 may be updated based on the actions performed using the pointer 420 .
- the visual search engine renders additional information about each object on the graphical user interface 200 .
- the metadata displays 430 provide information about the objects in the two-dimensional representation 440 .
- the metadata display 430 provides a limited amount of information that corresponds to the objects when the pointer 420 hovers over the objects in the visual representation.
- the visual search engine receives the limited amount of information from the objects database.
- the information is displayed on the graphical user interface 400 below the search box 410 when the user hovers over an object in the two-dimensional representation.
- the metadata display 430 may also be generated in a three-dimensional representation of the objects when the user hovers over an object.
- the two-dimensional representation 440 includes objects that satisfies the selections made by the user or the queries provided by the user.
- the visual search engine may generate the two-dimensional representation 440 when the set of objects is viewable within the display region of the graphical user interface without scrolling.
- the visual search engine traverses the objects database to obtain the images that are associated with each object in the set of objects.
- the visual search engine renders the images associated with each object in the two-dimensional representation 440 .
- the two-dimensional representation 440 includes object labels 450 .
- the visual search engine receives the information for each object label 450 from the objects database. In turn, the visual search engine renders a specific amount of data in the object label based on the number of objects that are displayed within the two-dimensional representation 440 . When the number of objects in the two-dimensional representation 440 covers more than a threshold amount of the total viewable region of the graphical user interface 400 , the visual search engine renders less information for each object. For instance, when the number of objects in the two-dimensional representation 440 covers more than 80 percent of the total viewable region of the graphical user interface 400 , the visual search engine may only render title information in the object label 450 for each object in the two-dimensional representation 440 . Otherwise, the visual explore engine, may render the title information, and the values for attribute frequently used to narrow a set of objects in the object label 450 .
- the two-dimensional representation generated by the visual search engine includes object labels that contain additional information for each object.
- the additional information comprises values associated with attributes associated with the objects that were not previously rendered by the visual explore engine.
- the additional information may be stored in the objects database.
- the additional information allows the user to determine whether the objects displayed in the two-dimensional representation satisfy unspecified user criteria associated with the additional information.
- the user may click on the object to initiate a subsequent action that helps the user complete the task or takes the user to additional information about the object. For instance, the click may initiate a search query corresponding to the object. If the object represents a consumer electronics product, e.g. digital camera, the click may initiate a purchase transaction on a purchase page corresponding to the camera.
- the click may initiate a search for accommodations and transportation to the destination.
- the user may continue searching by entering new search terms or selecting a category control that starts a new visual search.
- FIG. 5 is a graphical user interface 500 that illustrates a set of objects that match the criteria provided by category or refine controls in accordance with embodiments of the invention.
- the graphical user interface 500 includes a two-dimensional representation 510 and object labels 520 .
- the visual search engine may generate the two-dimensional representation 510 in graphical user interface 500 when the number of objects that are selected for display is viewable without scrolling.
- the two-dimensional representation 510 includes objects that satisfies the selections made by the user or the queries provided by the user.
- the visual search engine traverses the objects database to obtain the images that are associated with each object of the set of objects. In turn, the visual search engine renders the images associated with each object in the two-dimensional representation 510 .
- the two-dimensional representation 510 includes object labels 520 .
- the visual search engine receives the information for each object label 520 from the objects database. In turn, the visual search engine renders the values of attributes that are frequently used by users to reduce the set of objects in the object labels 520 .
- the objects selected by the visual search engine may be refined based on the context of the user interacting with the visual explore engine.
- the context of the user may include, but is not limited to, current user location, prior browsing history, or current configuration of the user's client device configuration.
- the visual search engine my use the context information to filter the objects selected from the objects database or the results provided by the index servers that correspond to an object selected by the user.
- an object in the two-dimensional representation is selected by the user.
- the visual search engine generates a graphical user interface that includes data from the object database and the URLs from the index servers.
- the graphical user interface displays the search results from the index server and object database to the user.
- FIG. 6 is a graphical user interface 600 that illustrates a result set 620 generated in response to click action in accordance with embodiments of the invention.
- the graphical user interface 600 includes a search box 610 , search results 620 , and a visual search link 630 .
- the visual search engine generates the graphical user interface 600 when the user clicks on an object with the pointer or selects an object.
- the search box 610 includes a search icon that causes the visual search engine to initiate a search with the search terms contained within the search box 610 .
- the search box 610 automatically displays the search query associated with the selected object. Because the user selected the object with the pointer, the visual search engine searches the index server and object database using the search query associated with the clicked object.
- the visual search engine updates the graphical user interface 600 with search results 620 .
- the search results 620 include an image associated with the selected object and values for a subset of the attributes that correspond to the selected object. In some embodiments, the values and corresponding subset of attributes are selected from the attributes that are frequently used by users to narrow the set of objects to the same selected object.
- the visual search engine may receive the image for the selected object and the corresponding attributes and values from the object database.
- the search results 620 also include URLs. The visual search engine receives the URLs from the index servers. The URLs point to web pages that contain the terms included in the search query of the selected object.
- the graphical user interface 600 includes a visual search link 630 .
- the visual search link 630 allows the user to begin a new visual search of a set of objects for an object that satisfies the user.
- the visual search engine when the user clicks on the visual explore link, the visual search engine generates a two-dimensional or three-dimensional representation having a previously generated set of objects.
- the visual search engine executes a method to visually search and explore a set of objects.
- the visual search engine receives a user query and determines whether to include a visual search link in a graphical user interface generated by the visual explore engine.
- the visual search link is prominently included in the graphical user interface when the user appears to be performing a search that is not directed to one particular object in the object database.
- the visual search link is less prominent.
- FIG. 7 is a logic diagram that illustrates a method to visually search through and explore a set of objects in accordance with embodiments of the invention.
- the computer-implemented method is initialized by the visual search engine in step 710 .
- the visual search engine is executed by a server that is communicatively connected to a client device.
- the server receives a query from the client device.
- the server determines whether to include a visual search link in a result set corresponding to the query.
- the server generates a result set that includes a list of uniform resource locators and the visual search link when the user query is general or specific.
- the results and the visual search link are transmitted from the server to the client device, in step 750 .
- the visual search link is configured to link to a three-dimensional interface or a two-dimensional interface having a collection of images corresponding to the set of objects that match the query.
- Each image is associated with a subsequent object query comprising metadata associated with the object corresponding to the image.
- the subsequent object query may be executed by the server in response to a click on an image corresponding to the object.
- the server may generate a results page having uniform resource locators (URLs) associated with web pages or documents that contain the terms of the subsequent query and information received from the objects database.
- URLs uniform resource locators
- the three-dimensional representation and the two-dimensional representation generated by the server may dynamically vary image dimensions in the collection images associated with the objects. Additionally, the three-dimensional representation and the two-dimensional representation generated by the server may dynamically vary metadata displayed in the three-dimensional representation and the two-dimensional representation.
- the server may generate a three-dimensional representation or a two-dimensional representation that dynamically varies controls included in the three-dimensional representation and the two-dimensional representation based on attributes of objects within the set of objects. The controls may refine, filter, categorize, or sort the set of objects. The method terminates in step 760 .
- the visual search engine identifies a match with a user query and selects a prominence assigned to the visual explore link.
- the match is a general match
- the visual search engine generates a very prominent visual search link to attract the user attention.
- position within the category control, color, size, or shape may be used to alter the prominence of the visual explore link.
- a higher position in the category control may correlate to very prominent and a lower position in the category control may correlate to low prominence.
- the visual search engine may identify a conceptual match, a conceptual subset match, and object level match.
- the conceptual matches cause the visual search engine to generate a very prominent visual search link to attract the user attention because the user search terms are very general.
- the conceptual matches may match the categories assigned to the objects.
- the conceptual subset match causes the visual search engine to generate a less prominent visual explore link.
- the conceptual subset matches may match the subcategories assigned to the objects.
- the object level match causes the visual search engine to generate the least prominent visual explore link.
- the object level match may match the actual object names assigned to the objects.
- the visual search engine generates synonyms for the user search terms based on definitions stored in a search log to identify the matches.
- embodiments of the invention allow a user to visually explore a set of objects and assists the user with formulating a query for a specific item.
- the graphical user interface generated by the visual search engine displays the set of objects.
- the graphical user interface may also display metadata that describes the objects when the user hovers over the object.
- the visual search engine may issue a query to the index server and objects database when the user selects an object in the graphical user interface.
Abstract
Methods, systems, graphical user interfaces, and computer-readable media for visually searching and exploring a set of objects are provided. A computer system executes a method that generates three-dimensional representations or two-dimensional representations for a set of objects in response to a user interaction with an interface that displays the three-dimensional representations or the two-dimensional representations. The interface includes filter controls, sorting controls, and classification controls, which are dynamically altered based on the content of a user query or the attributes of the objects in the three-dimensional representations or two-dimensional representations.
Description
- This patent application is a continuation of Attorney Docket No. 327771.01/MFCP.150920; U.S. application Ser. No. 12/551,236, filed 31 Aug. 2009, which is incorporated herein by reference in the entirety.
- Conventionally, search engines are configured to provide results that include one or more terms of a search query. Conventional search engines may use indices storing references to electronic documents and the terms included in the electronic documents to generate the results. The search engine includes the references to the electronic documents identified in an index having similar terms in the results.
- Some conventional search engines may provide search suggestions to a user when one or more terms are entered in a conventional search interface for the conventional search engine. The suggested terms may be selected and used to locate electronic documents that include the entered terms and the selected suggested terms. However, the terms entered in the search interface or the suggested terms may not correspond to an object that is sought by a user.
- The conventional search interface fails to provide adequate assistance in formulating terms for a set of objects that is sought by the user. A seeker hopes to complete a task upon locating the set of objects. But the seeker may be confronted with difficulty in formulating a search query that causes the search engine to provide the set of objects that is sought. For instance, the seeker may be searching for any of the following: best American-made hybrid cars on the market under $25,000; child-friendly movies playing today in my area; or netbooks under $400. The seeker may have difficulty locating the set of objects may resort to trial and error to formulate terms for the set of objects. Because the conventional search engine only performs term match using each term in the user search query, the actual content results often doesn't exactly correspond to the set of objects that is sought by the user.
- Embodiments of the invention overcoming these and other problems in the art relate in one regard to a computer system, graphical user interface, and computer-implemented method to visually search and explore a set of objects. The computer system renders images associated with a set of objects and transitions between a three-dimensional interface and two-dimensional interface when narrowing the set of objects.
- The computer system allows a user to visually search through and explore a set of objects related to a user's intent. The computer system includes a database and a server. The database is configured to store objects, attributes for each object, and images associated with each object. The server is connected to the database. The server retrieves a set of images and attributes in response to a visual search request. In turn, the server or client may generate a graphical user interface to organize the images and attributes. The graphical user interface may dynamically transition between a three-dimensional representation of the set of images and a two-dimensional representation of the set of images based on the number of images remaining in the set of images.
- This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the detailed description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used in isolation as an aid in determining the scope of the claimed subject matter.
-
FIG. 1 is a network diagram that illustrates an exemplary computing system in accordance with embodiments of the invention; -
FIG. 2 is a graphical user interface that illustrates a search results page having a visual search link in accordance with embodiments of the invention; -
FIG. 3 is a graphical user interface that illustrates images for a set of objects in a three-dimensional representation in accordance with embodiments of the invention; -
FIG. 4 is a graphical user interface that illustrates metadata surfaced for each object in a two-dimensional representation in accordance with embodiments of the invention; -
FIG. 5 is a graphical user interface that illustrates a set of objects that match the criteria provided by category or refine controls in accordance with embodiments of the invention; -
FIG. 6 is a graphical user interface that illustrates a result set generated in response to click action in accordance with embodiments of the invention; and -
FIG. 7 is a logic diagram that illustrates a method to visually search through and explore a set of objects in accordance with embodiments of the invention. - This patent describes the subject matter for patenting with specificity to meet statutory requirements. However, the description itself is not intended to limit the scope of this patent. Rather, the inventors have contemplated that the claimed subject matter might also be embodied in other ways, to include different steps or combinations of steps similar to the ones described in this patent, in conjunction with other present or future technologies. Moreover, although the terms “step” and “block” may be used herein to connote different elements of methods employed, the terms should not be interpreted as implying any particular order among or between various steps herein disclosed unless and except when the order of individual steps is explicitly described. Further, embodiments are described in detail below with reference to the attached drawing figures, which are incorporated in their entirety by reference herein.
- Embodiments of the invention provide a computer system that allows a user to visually search through and explore a set of objects included in a graphical user interface that contains the results to a user query or browsing request. The objects in the set of objects may correspond to, but is not limited to, shopping products, automobile products, animals, famous people, entertainment media, travel, or sports. In certain embodiments, the graphical user interface implements visual search and exploration by allowing the user to express ideas and search terms without having to formulate words for them. The graphical user interface may render a visual representation for each object and important attributes for the set of objects. The important attributes include attributes that users frequently use to evaluate the set of objects. Users can visually interact with the important attributes rather than formulate textual queries to narrow the set of objects. Additionally, the graphical user interface may include tangentially related categories to provide a user with an opportunity to serendipitously discover objects or relationships that are unknown to the user. Furthermore, each object in the graphical user interface may be linked with a query that may be previewed in a search box when a pointer hovers over the object. In turn, the query may be executed by clicking on the object. And the graphical user interface may dynamically adapt based on the number of items currently available in the set of objects. Accordingly, the computer system may be used to narrow a set of objects, to identify new items in a given field, e.g., new car models, new books, new movies in theaters—especially when the name of the objects are unknown, or to surface interesting and popular facts associated within a selected category.
- As one skilled in the art will appreciate, the computer system may include hardware, software, or a combination of hardware and software. The hardware includes processors and memories configured to execute instructions stored in the memories. In one embodiment, the memories include computer-readable media that store a computer-program product having computer-useable instructions for a computer-implemented method. Computer-readable media include both volatile and nonvolatile media, removable and nonremovable media, and media readable by a database, a switch, and various other network devices. Network switches, routers, and related components are conventional in nature, as are means of communicating with the same. By way of example, and not limitation, computer-readable media comprise computer-storage media and communications media. Computer-storage media, or machine-readable media, include media implemented in any method or technology for storing information. Examples of stored information include computer-useable instructions, data structures, program modules, and other data representations. Computer-storage media include, but are not limited to, random access memory (RAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), flash memory or other memory technology, compact-disc read only memory (CD-ROM), digital versatile discs (DVD), holographic media or other optical disc storage, magnetic cassettes, magnetic tape, magnetic disk storage, and other magnetic storage devices. These memory technologies can store data momentarily, temporarily, or permanently.
-
FIG. 1 is a network diagram that illustrates anexemplary operating environment 100, in accordance with embodiments of the invention. The operatingenvironment 100 shown inFIG. 1 is merely exemplary and is not intended to suggest any limitation as to scope or functionality. Embodiments of the invention are operable with numerous other configurations. With reference toFIG. 1 , the operatingenvironment 100 includes anetwork 110, avisual search engine 120,client devices 130,web pages 140, anobjects database 150, andindex servers 160. - The
network 110 is configured to facilitate communication between theclient devices 130 and thevisual search engine 120. Thenetwork 110 also facilitates communication between theobjects database 150, thevisual search engine 120, and theindex servers 160. Thenetwork 110 may be a communication network, such as a wireless network, local area network, wired network, or the Internet. In an embodiment, theclient devices 130 interact with thevisual search engine 120 utilizing thenetwork 110. In response, thevisual search engine 120 provides web pages, images, videos, or other electronic documents that contain terms provided or selected by the user. - The
visual search engine 120 is communicatively connected vianetwork 110 to theclient devices 130, theobjects database 150, and theindex servers 160. In certain embodiments, thevisual search engine 120 is a server that generates visual representations for display on theclient devices 130. Thevisual search engine 120 receives, overnetwork 110, selections of terms or selections of links fromclient devices 130 that provide interfaces that receive interactions from users. Thevisual search engine 120 may generate a three-dimensional or two-dimensional representation to display objects that correspond to terms or links selected by the user. The visual representation may include, but is not limited to, timelines, maps, grids or tables that organize the objects on the display. In one embodiment, thevisual search engine 120 is connected to a search engine to retrieve search results in response to a query formulated and issued by the visual explore engine. - In certain embodiments, the
visual search engine 120 traverses theobjects database 150 to identify objects that correspond to the selections received from the client devices. In turn, theobjects database 150 transmits a set of objects that satisfies the selections to thevisual search engine 120. The set of objects are associated with object queries,web pages 140, images, videos, or other electronic documents. - The
visual search engine 120 may also traverse theindex servers 160 to locateweb pages 140 that satisfy the object queries or user queries. In turn, theindex servers 160 transmit a set of uniform resource locators (URLs) that point to web pages, images, videos or other electronic documents that satisfy the object queries or user queries. Thevisual search engine 120 formats the URLs and transmits the URLs to theclient devices 130. - The
client devices 130 are utilized by a user to generate search terms, to hover over objects, or to select links or objects, and to receive results orweb pages 140 that are relevant to the search terms, the selected links, or the selected objects. Theclient devices 130 include, without limitation, personal digital assistants, smart phones, laptops, personal computers, gaming systems, set-top boxes, or any other suitable client computing device. Theclient devices 130 include user and system information storage to store user and system information on theclient devices 130. The user information may include search histories, cookies, and passwords. The system information may include internet protocol addresses, cached Web pages, and system utilization. Theclient devices 130 communicate with thevisual search engine 120 to receive the results orweb pages 140 that are relevant to the search terms, the selected links, or the selected objects. In some embodiments, theclient devices 130 execute a Silverlight browser application. The Silverlight browser application may provide an interactive experience when formulating search terms for an item and receiving results that response to a user request or the formulated search terms. When theclient device 130 receives objects orweb pages 140 selected by thevisual search engine 120, the Silverlight browser application only renders a subset of the items that satisfy the user request. Moreover, when the user scrolls through the display of theclient device 130, the Silverlight browser application may randomly populate the results with some of the additional objects by illustrating at least one object of the additional objects in each row of the display real estate of theclient device 130 as the visual search engine transmits the remaining additional objects to theclient device 130. - The
web pages 140 provide details about items that interest the user. Theweb pages 140 are indexed by theindex servers 160. Eachweb page 140 may include terms or metadata. The terms or metadata is used by the index servers to store the web page in an appropriate location. Additionally, theweb pages 140 are associated with URLs that are also stored by theindex servers 160. - The
objects database 150 stores attributes and images for each object. The attributes include titles, image size, image dimensions, and other metadata for the object. Thevisual search engine 120 may request one or more objects from theobjects database 150. In turn, theobjects database 150 transmits attributes and images to thevisual search engine 120. - The
index servers 160store web pages 140, terms associated with eachweb page 140, and uniform resource locators (URLs) corresponding to theweb pages 140. Thevisual search engine 120 may request one ormore web pages 140 from theindex servers 160. In turn, theindex servers 160 transmit the web pages to thevisual search engine 120. - Accordingly, the operating
environment 100 is configured with avisual search engine 120 that provides results that includeweb pages 140 and objects to theclient devices 130. Thevisual search engine 120 traverses theobject database 150 and traverses theindex servers 160 to obtain results that satisfy the requests received from the users. In turn, theclient devices 130 render the results for display to the users. - In an embodiment, the visual search engine generates a graphical user interface that includes results that match terms provided by or selections selected by the user. The results may include URLs that point to web pages. The graphical user interface contains a link that reformats the results into a visual representation. The link to the visual representation may be highlighted on the graphical user interface.
-
FIG. 2 is agraphical user interface 200 that illustrates a search results page having a visual search link in accordance with embodiments of the invention. Thegraphical user interface 200 includes asearch box 210 and avisual search link 230. - The
search box 210 in thegraphical user interface 200 allows the user to formulate their own search terms for a search query. Thesearch box 210 receives search terms provided by the user. In turn, the search terms are utilized by the visual search engine to obtain search results that match the terms entered in thesearch box 210. - The
visual search link 230 is part of thegraphical user interface 200. Thevisual search link 230 allows the user to explore results of a search query visually. Additionally, thevisual search link 230 may also allow the user to visually formulate a query without having to enter their own search terms. In some embodiments, the visual representation of the search results is a three-dimensional representation. - For instance, when a user enters “cars” in the
search box 210 and initiates a search using the term “cars” the visual search engine may recognize that the user hopes to complete a task associated with purchasing a used or new car. The results that a search engine returns may not aid the user in completing the task associated with purchasing a used or new car. The visual search engine ignores the results from the search engine and provides a visual display having visual representations of cars that are currently for sale in the user's location. The user may interact with the visual representation and features of thegraphical user interface 200 to locate the “cars” that the user seeks. - Alternatively, when a user enters “cars” in the
search box 210 and initiates a search using the term “cars” the visual search engine may return results that consist ofweb pages 140 that include the term car. However, when the user clicks on thevisual search link 230, the visual search engine generates a visual representation of the results. The user may interact with the visual representation and features of thegraphical user interface 200 to locate the “cars” that the user seeks. - In certain embodiments, the visual search engine may provide a browse entry point that does not require entry of a query terms. The browse entry page may include a collection of objects arranged by category. Each of the objects may be represented by an image and may correspond to a visual query.
- In one embodiment, the visual representation generated by the visual explore engine, in response to a user request, to visually explore the search results is a three-dimensional representation of the search results. A graphical user interface displays the three-dimensional representation and controls that are used to interact with the three-dimensional representation. In turn, the user can narrow the objects included in the three-dimensional representation.
-
FIG. 3 is agraphical user interface 300 that illustrates images for a set ofobjects 350 a in a three-dimensional representation 350 in accordance with embodiments of the invention. The visual search engine generates a visual representation ingraphical user interface 300 when the user clicks on the visual search link in a graphical user interface that includes URL results for search terms provided by the user. Thegraphical user interface 300 replaces the URL results with the visual representation (three-dimensional or two-dimensional) based on the number ofobjects 350 a and the dimensions of theobjects 350 a. The visual search engine animates the objects as the objects are rendered in thegraphical user interface 300. When the visual search engine generates the three-dimensional representation 350, theobjects 350 a move from the background of thegraphical user interface 300 and arrange themselves into an overlapping three-dimensional array of objects. In turn, the visual search engine animates controls on the left and right of thegraphical user interface 300 by floating the controls onto thegraphical user interface 300 from the left and right. The user can move a pointer to interact with the controls that that are on the left side and the right side of thegraphical user interface 300 to manipulate theobjects 350 a. In some embodiments, the visual search engine generates the visual representation in thegraphical user interface 300 when the user clicks on a visual search link without requiring the user to provide search terms. - The
graphical user interface 300 includescategory control 310,count control 320, refinecontrols 330,sort control 340, andscroll control 360 that are used to interact with theobjects 350 a included in the three-dimensional representation 350. Thegraphical user interface 300 allows the users to view all objects that match the search terms provided by the users. Thegraphical user interface 300 loads less then all of the objects but thescroll control 360 provides an infinite scroll feature that allows the users to view additional objects in the visual representation that are currently outside of the viewable region of the three-dimensional representation 350. In one embodiment, thegraphical user interface 300 may load the data for all objects, but stream the images as the specific objects come into view as thescroll control 360 reveals the additional objects. In another embodiment, the graphical user interface may only load a subset of the data and the corresponding images to allow the user to quickly interact with some of the objects. The additional data and images are loaded as requested via thescroll control 360. - The
category control 310 identifies the categories associated with each object in the three-dimensional representation 350. The number objects 350 a in the three-dimensional representation 350 may be reduced when the user selects thecategory control 310. In some embodiments, the three-dimensional representation 350 animates theobjects 350 a as the visual search engine reduces the number ofobjects 350 a in the three-dimensional representation 350. Theobjects 350 a that are not within the selected category drop or fade into a background of thegraphical user interface 300.Additional objects 350 a that satisfy the selected category replace the removedobjects 350 a in the three-dimensional representation 350. All theobjects 350 a that are in the updated three-dimensional representation 350 satisfy a category selected in thecategory control 310. - For instance, a user may select sport utility vehicle (SUV) as a category when searching for cars. In response to the user selection, the visual search engine updates the
objects 350 a in the three-dimensional representation 350 by removingobjects 350 a that are not categorized as SUVs. Theobjects 350 a that do not satisfy the user selection are animated by the visual search engine to fall out of the three-dimensional representation 350 and additional objects that are outside the viewable region of the three-dimensional representation 350 are animated to replace the removedobjects 350 a. - In another embodiment, the visual search engine may update the
category control 310 with a new link based on a previous search performed by the user. When the user's previous search is for a specific item or a branded item, the visual search engine creates a link that is associated with a visual representation of objects having features similar to the specific item or branded item. When the user selects the link, the visual search engine generates an interface having objects with similar features. Additionally, the visual search engine may include links in thecategory control 310 to information that is tangentially related to the previous search. For instance, if a user was looking for all of the new cars for sale and hovered over an Aston Martin, thecategory control 310 may update to show links to James Bond movies currently playing in movie theaters (because James Bond drives Aston Martins), and the country England (because Aston Martins are produced in England). - The
count control 320 displays the total number of items that are in the visual representation and the number of items that are currently displayed in the viewable regions of thegraphical user interface 300. Thecount control 320 updates as the user reduces or increases the objects that are within the visual representation (e.g. three-dimensional representation 350). Thecount control 320 updates when the user initiates a search with new or additional search terms. Alternatively, thecount control 320 updates when the user interacts with thecategory control 310 or refinecontrols 330 to manipulate theobjects 350 a in thegraphical user interface 300. - The refine
controls 330 represent attributes that are shared by some of theobjects 350 a that are within the visual representation. The number objects 350 a in the three-dimensional representation 350 may be reduced when the user selects a value for the attributes represented by the refine controls 330. In some embodiments, the three-dimensional representation 350 animates theobjects 350 a as the visual search engine reduces the number ofobjects 350 a in the three-dimensional representation 350. Theobjects 350 a that fail to satisfy the value selected for the attributes represented by the refinecontrols 330 drop or fade into a background of thegraphical user interface 300.Additional objects 350 a that satisfy the value selected for the attributes represented by the refinecontrols 330 replace the removedobjects 350 a. All theobjects 350 a that are in the updated visual representation (e.g. updated three-dimensional representation 350) satisfy the values selected for the attributes represented by the refine controls 330. In some embodiments, the attributes in the refinecontrols 350 are updated to provide only valid filtering options. For instance, when a user selects “Sport Utility Vehicle” as the vehicle type filter, the brand filter may automatically to remove manufacturers or brands that do not offer a “Sport Utility Vehicle.” - The
sort control 340 allows the user to sort theobjects 350 a. Thesort control 340 reorganizes theobjects 350 a in the three-dimensional representation 350. Thesort control 330 represents attributes that are shared by all theobjects 350 a in the visual representation. Theobjects 350 a are sorted alphabetically by default. When the user selects the sort control, the visual search engine can reorganize theobjects 350 a in the visual representation. For instance, when the user is searching for cars, thesort control 340 may allow the user to sort theobjects 350 a by brand, safety rating, or expert rating. The visual search engine obtains the brand, safety rating, and expert rating from the objects database, which stores attributes for brand, safety rating, and expert rating for the objects. After the user selects an attribute, the visual search engine reorders the objects in the visual representation. In an embodiment, the attributes that are included in sort control are selected based on search frequency in a search log. - The
scroll control 360 allows the user to viewadditional objects 350 a that are within the visual representation. In some embodiments, the scroll control provides infinite scroll, which allows thegraphical user interface 300 to continuously scroll through theobjects 350 a in the visual representation. The infinite scroll continuously loops through allobjects 350 a within the visual representation. - In some embodiments, the visual search engine transitions between a three-
dimensional representation 350 of theobjects 350 a and a two-dimensional representation. The visual search engine generates the threedimensional representation 350 when the number of objects that are selected for display in the graphical user interface is large. When the user reduces the number of objects to a reasonable number, the visual search engine may render the objects in a two-dimensional representation. In one embodiment, the remaining objects are reoriented by the visual search engine in rows that are viewable without scrolling. Also, the visual search engine may render a boundary around each of the remainingobjects 350 a. - For instance, the visual search engine may set a display default that requires generating a graphical user interface with detailed information for each object. However, the graphical user interface must also display as many of the objects that satisfy the user selection in the visual representation. The three-dimensional
visual representation 350 a of the objects may be generated by the visual search engine when the number of objects is large and showing detailed information would obscure the graphical user interface. The two-dimensional visual representation may be generated by the visual search engine when the number of objects is small and showing detailed information would not obscure the graphical user interface. Additionally, because the available display real estate increases, the visual search engine displays more information for each object. In some embodiments, images associated with each object are sized to 100 pixel by 100 pixels and the two-dimensional representation includes an array having a minimum of five rows and five columns of images. - In an embodiment, a two-dimensional representation of the objects is generated by the visual explore engine. The two-dimensional representation allows the user to view additional information about each object on the graphical user interface. The visual search engine uses the available display real estate on the graphical user interface to maximize information that is rendered for display to the user.
-
FIG. 4 is agraphical user interface 400 that illustrates metadata surfaced for each object in a two-dimensional representation 440 in accordance with embodiments of the invention. The visual search engine generates the two-dimensional representation 440 ingraphical user interface 400 based on the number of objects that are selected for rendering. In one embodiment, the visual search engine may generate the two-dimensional representation 440 ingraphical user interface 400 when the number of objects that are selected for display is viewable without scrolling. In some embodiments, the visual search engine may reformat the images associated with the objects to fit in thegraphical user interface 400 without scrolling. In other embodiments, the visual search engine may generate the two-dimensional representation 440 ingraphical user interface 400 when the number of objects allows for comparison of the objects based on the available metadata that is viewable on thegraphical user interface 400. - The
graphical user interface 400 includessearch box 410,pointer 420, metadata displays 430, the two-dimensional representation 440, and object labels 450. Thesearch box 410 allows the user to enter search terms that should be sent to the visual explore engine. Thesearch box 410 includes a search icon that causes the visual search engine to initiate a search with the search terms contained within thesearch box 410. Thesearch box 410 also automatically displays search queries associated with each object in the two-dimensional representation 440 when a user hovers over an object with thepointer 420. The user may initiate a search using the search queries that are automatically displayed in thesearch box 410 in response to the hover by clicking on the search icon included in thesearch box 410. Alternatively, the user may initiate a search by using the search queries associated with each object by clicking on the object with thepointer 420. When the user clicks on the object with thepointer 420, the visual search engine searches the index server and objects database using the search queries associated with the object. - In some embodiments, the
pointer 420 is a mouse that allows the user to hover over objects, click on objects, or drag and drop objects. Thegraphical user interface 400 may be updated based on the actions performed using thepointer 420. When the pointer hovers over an object in the two-dimensional representation 440, the visual search engine renders additional information about each object on thegraphical user interface 200. - The metadata displays 430 provide information about the objects in the two-
dimensional representation 440. In one embodiment, themetadata display 430 provides a limited amount of information that corresponds to the objects when thepointer 420 hovers over the objects in the visual representation. The visual search engine receives the limited amount of information from the objects database. In certain embodiments, the information is displayed on thegraphical user interface 400 below thesearch box 410 when the user hovers over an object in the two-dimensional representation. In another embodiment, themetadata display 430 may also be generated in a three-dimensional representation of the objects when the user hovers over an object. - The two-
dimensional representation 440 includes objects that satisfies the selections made by the user or the queries provided by the user. The visual search engine may generate the two-dimensional representation 440 when the set of objects is viewable within the display region of the graphical user interface without scrolling. The visual search engine traverses the objects database to obtain the images that are associated with each object in the set of objects. In turn, the visual search engine renders the images associated with each object in the two-dimensional representation 440. - In certain embodiments, the two-
dimensional representation 440 includes object labels 450. The visual search engine receives the information for eachobject label 450 from the objects database. In turn, the visual search engine renders a specific amount of data in the object label based on the number of objects that are displayed within the two-dimensional representation 440. When the number of objects in the two-dimensional representation 440 covers more than a threshold amount of the total viewable region of thegraphical user interface 400, the visual search engine renders less information for each object. For instance, when the number of objects in the two-dimensional representation 440 covers more than 80 percent of the total viewable region of thegraphical user interface 400, the visual search engine may only render title information in theobject label 450 for each object in the two-dimensional representation 440. Otherwise, the visual explore engine, may render the title information, and the values for attribute frequently used to narrow a set of objects in theobject label 450. - In another embodiment, the two-dimensional representation generated by the visual search engine includes object labels that contain additional information for each object. The additional information comprises values associated with attributes associated with the objects that were not previously rendered by the visual explore engine. The additional information may be stored in the objects database. The additional information allows the user to determine whether the objects displayed in the two-dimensional representation satisfy unspecified user criteria associated with the additional information. When the objects in the two-dimensional representation satisfy the user, the user may click on the object to initiate a subsequent action that helps the user complete the task or takes the user to additional information about the object. For instance, the click may initiate a search query corresponding to the object. If the object represents a consumer electronics product, e.g. digital camera, the click may initiate a purchase transaction on a purchase page corresponding to the camera. If the object represents a travel product, e.g., travel to a destination city, monument, or memorial, the click may initiate a search for accommodations and transportation to the destination. Alternatively, when the objects in the two-dimensional representation fail to satisfy the user, the user may continue searching by entering new search terms or selecting a category control that starts a new visual search.
-
FIG. 5 is agraphical user interface 500 that illustrates a set of objects that match the criteria provided by category or refine controls in accordance with embodiments of the invention. Thegraphical user interface 500 includes a two-dimensional representation 510 and object labels 520. - The visual search engine may generate the two-
dimensional representation 510 ingraphical user interface 500 when the number of objects that are selected for display is viewable without scrolling. The two-dimensional representation 510 includes objects that satisfies the selections made by the user or the queries provided by the user. The visual search engine traverses the objects database to obtain the images that are associated with each object of the set of objects. In turn, the visual search engine renders the images associated with each object in the two-dimensional representation 510. - In certain embodiments, the two-
dimensional representation 510 includes object labels 520. The visual search engine receives the information for eachobject label 520 from the objects database. In turn, the visual search engine renders the values of attributes that are frequently used by users to reduce the set of objects in the object labels 520. - The objects selected by the visual search engine may be refined based on the context of the user interacting with the visual explore engine. The context of the user may include, but is not limited to, current user location, prior browsing history, or current configuration of the user's client device configuration. The visual search engine my use the context information to filter the objects selected from the objects database or the results provided by the index servers that correspond to an object selected by the user.
- In some embodiments, an object in the two-dimensional representation is selected by the user. In turn, the visual search engine generates a graphical user interface that includes data from the object database and the URLs from the index servers. The graphical user interface displays the search results from the index server and object database to the user.
-
FIG. 6 is agraphical user interface 600 that illustrates a result set 620 generated in response to click action in accordance with embodiments of the invention. Thegraphical user interface 600 includes asearch box 610, search results 620, and avisual search link 630. - The visual search engine generates the
graphical user interface 600 when the user clicks on an object with the pointer or selects an object. Thesearch box 610 includes a search icon that causes the visual search engine to initiate a search with the search terms contained within thesearch box 610. Thesearch box 610 automatically displays the search query associated with the selected object. Because the user selected the object with the pointer, the visual search engine searches the index server and object database using the search query associated with the clicked object. - In turn, the visual search engine updates the
graphical user interface 600 with search results 620. The search results 620 include an image associated with the selected object and values for a subset of the attributes that correspond to the selected object. In some embodiments, the values and corresponding subset of attributes are selected from the attributes that are frequently used by users to narrow the set of objects to the same selected object. The visual search engine may receive the image for the selected object and the corresponding attributes and values from the object database. The search results 620 also include URLs. The visual search engine receives the URLs from the index servers. The URLs point to web pages that contain the terms included in the search query of the selected object. - The
graphical user interface 600 includes avisual search link 630. Thevisual search link 630 allows the user to begin a new visual search of a set of objects for an object that satisfies the user. In some embodiments, when the user clicks on the visual explore link, the visual search engine generates a two-dimensional or three-dimensional representation having a previously generated set of objects. - In some embodiments, the visual search engine executes a method to visually search and explore a set of objects. The visual search engine receives a user query and determines whether to include a visual search link in a graphical user interface generated by the visual explore engine. The visual search link is prominently included in the graphical user interface when the user appears to be performing a search that is not directed to one particular object in the object database. When the user is performing a search that is directed to a particular object, the visual search link is less prominent.
-
FIG. 7 is a logic diagram that illustrates a method to visually search through and explore a set of objects in accordance with embodiments of the invention. The computer-implemented method is initialized by the visual search engine instep 710. The visual search engine is executed by a server that is communicatively connected to a client device. Instep 720, the server receives a query from the client device. Instep 730, the server determines whether to include a visual search link in a result set corresponding to the query. Instep 740, the server generates a result set that includes a list of uniform resource locators and the visual search link when the user query is general or specific. The results and the visual search link are transmitted from the server to the client device, instep 750. - The visual search link is configured to link to a three-dimensional interface or a two-dimensional interface having a collection of images corresponding to the set of objects that match the query. Each image is associated with a subsequent object query comprising metadata associated with the object corresponding to the image. The subsequent object query may be executed by the server in response to a click on an image corresponding to the object. When the server receives the subsequent object query, the server may generate a results page having uniform resource locators (URLs) associated with web pages or documents that contain the terms of the subsequent query and information received from the objects database.
- The three-dimensional representation and the two-dimensional representation generated by the server may dynamically vary image dimensions in the collection images associated with the objects. Additionally, the three-dimensional representation and the two-dimensional representation generated by the server may dynamically vary metadata displayed in the three-dimensional representation and the two-dimensional representation. The server may generate a three-dimensional representation or a two-dimensional representation that dynamically varies controls included in the three-dimensional representation and the two-dimensional representation based on attributes of objects within the set of objects. The controls may refine, filter, categorize, or sort the set of objects. The method terminates in
step 760. - In an embodiment, the visual search engine identifies a match with a user query and selects a prominence assigned to the visual explore link. When the match is a general match, the visual search engine generates a very prominent visual search link to attract the user attention. For instance, position within the category control, color, size, or shape may be used to alter the prominence of the visual explore link. A higher position in the category control may correlate to very prominent and a lower position in the category control may correlate to low prominence.
- In some embodiments, the visual search engine may identify a conceptual match, a conceptual subset match, and object level match. The conceptual matches cause the visual search engine to generate a very prominent visual search link to attract the user attention because the user search terms are very general. The conceptual matches may match the categories assigned to the objects. The conceptual subset match causes the visual search engine to generate a less prominent visual explore link. The conceptual subset matches may match the subcategories assigned to the objects. The object level match causes the visual search engine to generate the least prominent visual explore link. The object level match may match the actual object names assigned to the objects. In certain embodiments, the visual search engine generates synonyms for the user search terms based on definitions stored in a search log to identify the matches.
- In summary, embodiments of the invention allow a user to visually explore a set of objects and assists the user with formulating a query for a specific item. The graphical user interface generated by the visual search engine displays the set of objects. The graphical user interface may also display metadata that describes the objects when the user hovers over the object. Additionally, the visual search engine may issue a query to the index server and objects database when the user selects an object in the graphical user interface.
- The foregoing descriptions of the embodiments of the invention are illustrative, and modifications in configuration and implementation are within the scope of the current description. For instance, while the embodiments of the invention are generally described with relation to
FIGS. 1-7 , those descriptions are exemplary. Although the subject matter has been described in language specific to structural features or methodological acts, it is understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims. The scope of the embodiment of the invention is accordingly intended to be limited only by the following claims.
Claims (20)
1. A computer-implemented method to visually search and explore a set of objects, the computer-implemented method comprising:
receiving, by a server, a query;
determining, by the server, whether to include a visual search link in a result set corresponding to the query;
generating a result set that includes a list of uniform resource locators and the visual search link when the user query is general or specific; and
transmitting the results and the visual search link to the user.
2. The computer-implemented method of claim 1 , wherein the visual search link is configured to link to a three-dimensional interface or a two-dimensional interface having a collection of images corresponding to the set of objects that match the query.
3. The computer-implemented method of claim 2 , wherein the three-dimensional interface and the two-dimensional interface dynamically vary image sizes in the collection of images.
4. The computer-implemented method of claim 3 , wherein the three-dimensional interface and the two-dimensional interface dynamically vary metadata displayed in the three-dimensional interface and the two-dimensional interface.
5. The computer-implemented method of claim 4 , wherein the three-dimensional interface and the two-dimensional interface dynamically vary controls included in the three-dimensional interface and the two-dimensional interface based on attributes of objects within the set of objects.
6. The computer-implemented method of claim 5 , wherein the controls refine, filter, categorize, or sort the set of objects.
7. The computer-implemented method of claim 2 , wherein each image is associated with a subsequent query comprising metadata associated with the object corresponding to the image.
8. The computer-implemented method of claim 7 , wherein the subsequent query includes context information corresponding to the user and is executed in response to a click on the image.
9. The computer-implemented method of claim 6 , further comprising: receiving the subsequent query and generating a results page having uniform resource locators (URLS) associated with web pages or documents that contain the terms of the subsequent query.
10. The computer-implemented method of claim 7 , wherein the metadata is displayed on the three-dimensional interface or the two-dimensional interface when the image is subject to a hover action.
11. A computing device that executes instruction to present a graphical user interface configured to visually search and explore a set of objects, the graphical user interface, comprising:
a sorting control configured to reorganize the set of objects based on attributes shared by all of the objects within the set of objects;
a filtering control configured to reduce the number of objects in the set of objects based on the attributes of each object in the set of objects;
a classification control configured to reduce the number of objects in the set based on categories represented within the set of objects; and
the graphical user interface is configured to display images of objects that match a user query and to transition between a three-dimensional view and a two-dimensional view based on the number of objects that are within the set of objects.
12. The device of claim 11 , wherein the attributes include items that are frequently requested by a number of users.
13. The device of claim 11 , wherein the attributes include movie items and country of origin items that are related to the query.
14. The device of claim 11 , further comprising: a search box control that is automatically populated with a subsequent query corresponding to an object that is a subject of a hover action.
15. The device of claim 14 , wherein the subsequent query is executed when the object is a subject of a click action.
16. A computer-storage media having computer-executable instructions for performing a computer-implemented method to visually search and explore a set of objects, the computer-implemented method comprising:
receiving a query;
generating a result set that includes a list of uniform resource locators and a visual search link; and
transmitting the results and the visual search link to the user, wherein the visual search link is part of the result set having a set of images and attributes, wherein the set of images is rendered in a three-dimensional interface or a two-dimensional interface based on a number of images that are within the set of images:
the two-dimensional interface is provided if the number of images is below a threshold and the user accesses the visual search link, and
the three-dimensional interface is provided if the number of images is above the threshold and the user accesses the visual search link.
17. The media of claim 16 , wherein the three-dimensional interface and the two-dimensional interface dynamically vary image sizes in the collection of images.
18. The media claim 16 , wherein the three-dimensional interface and the two-dimensional interface dynamically vary metadata displayed in the three-dimensional interface and the two-dimensional interface.
19. The media of claim 16 , wherein a prominence of the visual search link varies based on a number images or a number of uniform resource locators in the set of results.
20. The media of claim 16 , wherein controls included in the three-dimensional interface or the two-dimensional interface are modified dynamically based on attributes of images within the set of images.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/690,472 US20130091471A1 (en) | 2009-08-31 | 2012-11-30 | Visual search and three-dimensional results |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/551,236 US8335784B2 (en) | 2009-08-31 | 2009-08-31 | Visual search and three-dimensional results |
US13/690,472 US20130091471A1 (en) | 2009-08-31 | 2012-11-30 | Visual search and three-dimensional results |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/551,236 Continuation US8335784B2 (en) | 2009-08-31 | 2009-08-31 | Visual search and three-dimensional results |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130091471A1 true US20130091471A1 (en) | 2013-04-11 |
Family
ID=43626366
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/551,236 Active 2030-09-11 US8335784B2 (en) | 2009-08-31 | 2009-08-31 | Visual search and three-dimensional results |
US13/690,472 Abandoned US20130091471A1 (en) | 2009-08-31 | 2012-11-30 | Visual search and three-dimensional results |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/551,236 Active 2030-09-11 US8335784B2 (en) | 2009-08-31 | 2009-08-31 | Visual search and three-dimensional results |
Country Status (11)
Country | Link |
---|---|
US (2) | US8335784B2 (en) |
EP (1) | EP2473934A4 (en) |
JP (1) | JP5320509B2 (en) |
KR (2) | KR101691249B1 (en) |
CN (1) | CN102483760B (en) |
AU (1) | AU2010286753B2 (en) |
BR (1) | BR112012004517A2 (en) |
CA (1) | CA2768966C (en) |
SG (2) | SG10201404668VA (en) |
TW (1) | TWI490714B (en) |
WO (1) | WO2011025749A2 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8954544B2 (en) | 2010-09-30 | 2015-02-10 | Axcient, Inc. | Cloud-based virtual machines and offices |
US9104621B1 (en) | 2010-09-30 | 2015-08-11 | Axcient, Inc. | Systems and methods for restoring a file |
US9213607B2 (en) | 2010-09-30 | 2015-12-15 | Axcient, Inc. | Systems, methods, and media for synthesizing views of file system backups |
US9235474B1 (en) | 2011-02-17 | 2016-01-12 | Axcient, Inc. | Systems and methods for maintaining a virtual failover volume of a target computing system |
US9292153B1 (en) * | 2013-03-07 | 2016-03-22 | Axcient, Inc. | Systems and methods for providing efficient and focused visualization of data |
US9397907B1 (en) | 2013-03-07 | 2016-07-19 | Axcient, Inc. | Protection status determinations for computing devices |
US9705730B1 (en) | 2013-05-07 | 2017-07-11 | Axcient, Inc. | Cloud storage using Merkle trees |
US9785647B1 (en) | 2012-10-02 | 2017-10-10 | Axcient, Inc. | File system virtualization |
US9852140B1 (en) | 2012-11-07 | 2017-12-26 | Axcient, Inc. | Efficient file replication |
US10055092B2 (en) | 2014-05-19 | 2018-08-21 | Samsung Electronics Co., Ltd. | Electronic device and method of displaying object |
US10284437B2 (en) | 2010-09-30 | 2019-05-07 | Efolder, Inc. | Cloud-based virtual machines and offices |
US10437423B2 (en) | 2015-10-30 | 2019-10-08 | Hyland Software, Inc. | Methods and apparatuses for providing an infinitely scrolling accumulator |
US10838611B1 (en) * | 2019-05-03 | 2020-11-17 | Google Llc | Navigation system on the infinite scroll |
US20220147531A1 (en) * | 2018-10-29 | 2022-05-12 | State Farm Mutual Automobile Insurance Company | Dynamic data-driven consolidation of user interface interactions requesting roadside assistance |
US20220207103A1 (en) * | 2014-02-12 | 2022-06-30 | Pinterest, Inc. | Visual search refinement |
Families Citing this family (92)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7509588B2 (en) | 2005-12-30 | 2009-03-24 | Apple Inc. | Portable electronic device with interface reconfiguration mode |
US10313505B2 (en) | 2006-09-06 | 2019-06-04 | Apple Inc. | Portable multifunction device, method, and graphical user interface for configuring and displaying widgets |
US8519964B2 (en) | 2007-01-07 | 2013-08-27 | Apple Inc. | Portable multifunction device, method, and graphical user interface supporting user navigations of graphical objects on a touch screen display |
US8619038B2 (en) | 2007-09-04 | 2013-12-31 | Apple Inc. | Editing interface |
US8549407B2 (en) * | 2007-12-05 | 2013-10-01 | Ebay Inc. | Multi-dimensional dynamic visual browsing |
US9069854B2 (en) * | 2008-10-18 | 2015-06-30 | Pomian & Corella, Llc | Facilitating browsing of result sets |
EP2341450A1 (en) * | 2009-08-21 | 2011-07-06 | Mikko Kalervo Väänänen | Method and means for data searching and language translation |
US9501519B1 (en) | 2009-12-14 | 2016-11-22 | Amazon Technologies, Inc. | Graphical item chooser |
US10007393B2 (en) * | 2010-01-19 | 2018-06-26 | Apple Inc. | 3D view of file structure |
US9792638B2 (en) * | 2010-03-29 | 2017-10-17 | Ebay Inc. | Using silhouette images to reduce product selection error in an e-commerce environment |
US10788976B2 (en) | 2010-04-07 | 2020-09-29 | Apple Inc. | Device, method, and graphical user interface for managing folders with multiple pages |
US8423911B2 (en) | 2010-04-07 | 2013-04-16 | Apple Inc. | Device, method, and graphical user interface for managing folders |
US8266542B2 (en) * | 2010-06-17 | 2012-09-11 | Hewlett-Packard Development Company, L.P. | Rendering and sorting book search results as three dimensional icons |
US9497084B2 (en) * | 2011-02-28 | 2016-11-15 | Ricoh Company, Ltd. | Display filtering in device management system |
US20120290985A1 (en) * | 2011-05-11 | 2012-11-15 | Barnes & Noble, Inc. | System and method for presenting and interacting with eperiodical subscriptions |
US9390137B2 (en) | 2011-06-29 | 2016-07-12 | Nokia Technologies Oy | Method and apparatus for providing an ordering metric for a multi-dimensional contextual query |
US8706860B2 (en) | 2011-06-30 | 2014-04-22 | Amazon Technologies, Inc. | Remote browsing session management |
US8799412B2 (en) | 2011-06-30 | 2014-08-05 | Amazon Technologies, Inc. | Remote browsing session management |
US8577963B2 (en) | 2011-06-30 | 2013-11-05 | Amazon Technologies, Inc. | Remote browsing session between client browser and network based browser |
US9621406B2 (en) | 2011-06-30 | 2017-04-11 | Amazon Technologies, Inc. | Remote browsing session management |
US9037696B2 (en) | 2011-08-16 | 2015-05-19 | Amazon Technologies, Inc. | Managing information associated with network resources |
US9195768B2 (en) | 2011-08-26 | 2015-11-24 | Amazon Technologies, Inc. | Remote browsing session management |
US10089403B1 (en) | 2011-08-31 | 2018-10-02 | Amazon Technologies, Inc. | Managing network based storage |
US10693991B1 (en) | 2011-09-27 | 2020-06-23 | Amazon Technologies, Inc. | Remote browsing session management |
US8914514B1 (en) | 2011-09-27 | 2014-12-16 | Amazon Technologies, Inc. | Managing network based content |
US9178955B1 (en) | 2011-09-27 | 2015-11-03 | Amazon Technologies, Inc. | Managing network based content |
US8849802B2 (en) | 2011-09-27 | 2014-09-30 | Amazon Technologies, Inc. | Historical browsing session management |
US9298843B1 (en) | 2011-09-27 | 2016-03-29 | Amazon Technologies, Inc. | User agent information management |
US9152970B1 (en) | 2011-09-27 | 2015-10-06 | Amazon Technologies, Inc. | Remote co-browsing session management |
US9641637B1 (en) | 2011-09-27 | 2017-05-02 | Amazon Technologies, Inc. | Network resource optimization |
US9383958B1 (en) | 2011-09-27 | 2016-07-05 | Amazon Technologies, Inc. | Remote co-browsing session management |
US8589385B2 (en) | 2011-09-27 | 2013-11-19 | Amazon Technologies, Inc. | Historical browsing session management |
US8615431B1 (en) | 2011-09-29 | 2013-12-24 | Amazon Technologies, Inc. | Network content message placement management |
US9313100B1 (en) | 2011-11-14 | 2016-04-12 | Amazon Technologies, Inc. | Remote browsing session management |
JP6278893B2 (en) * | 2011-11-24 | 2018-02-14 | マイクロソフト テクノロジー ライセンシング,エルエルシー | Interactive multi-mode image search |
US8972477B1 (en) | 2011-12-01 | 2015-03-03 | Amazon Technologies, Inc. | Offline browsing session management |
US9117002B1 (en) | 2011-12-09 | 2015-08-25 | Amazon Technologies, Inc. | Remote browsing session management |
US9009334B1 (en) | 2011-12-09 | 2015-04-14 | Amazon Technologies, Inc. | Remote browsing session management |
US9330188B1 (en) | 2011-12-22 | 2016-05-03 | Amazon Technologies, Inc. | Shared browsing sessions |
US8924890B2 (en) * | 2012-01-10 | 2014-12-30 | At&T Intellectual Property I, L.P. | Dynamic glyph-based search |
US8429163B1 (en) | 2012-01-25 | 2013-04-23 | Hewlett-Packard Development Company, L.P. | Content similarity pyramid |
US9087024B1 (en) | 2012-01-26 | 2015-07-21 | Amazon Technologies, Inc. | Narration of network content |
US8627195B1 (en) | 2012-01-26 | 2014-01-07 | Amazon Technologies, Inc. | Remote browsing and searching |
US8839087B1 (en) * | 2012-01-26 | 2014-09-16 | Amazon Technologies, Inc. | Remote browsing and searching |
US9509783B1 (en) | 2012-01-26 | 2016-11-29 | Amazon Technlogogies, Inc. | Customized browser images |
US9092405B1 (en) | 2012-01-26 | 2015-07-28 | Amazon Technologies, Inc. | Remote browsing and searching |
US9336321B1 (en) | 2012-01-26 | 2016-05-10 | Amazon Technologies, Inc. | Remote browsing and searching |
US9037975B1 (en) | 2012-02-10 | 2015-05-19 | Amazon Technologies, Inc. | Zooming interaction tracking and popularity determination |
US9183258B1 (en) | 2012-02-10 | 2015-11-10 | Amazon Technologies, Inc. | Behavior based processing of content |
US9137210B1 (en) | 2012-02-21 | 2015-09-15 | Amazon Technologies, Inc. | Remote browsing session management |
US8745058B1 (en) | 2012-02-21 | 2014-06-03 | Google Inc. | Dynamic data item searching |
US9374244B1 (en) | 2012-02-27 | 2016-06-21 | Amazon Technologies, Inc. | Remote browsing session management |
US10296558B1 (en) | 2012-02-27 | 2019-05-21 | Amazon Technologies, Inc. | Remote generation of composite content pages |
US9208316B1 (en) | 2012-02-27 | 2015-12-08 | Amazon Technologies, Inc. | Selective disabling of content portions |
US9460220B1 (en) | 2012-03-26 | 2016-10-04 | Amazon Technologies, Inc. | Content selection based on target device characteristics |
US9307004B1 (en) | 2012-03-28 | 2016-04-05 | Amazon Technologies, Inc. | Prioritized content transmission |
CN102693318B (en) * | 2012-05-30 | 2014-06-11 | 华为技术有限公司 | Report query method and report query equipment |
US9098516B2 (en) * | 2012-07-18 | 2015-08-04 | DS Zodiac, Inc. | Multi-dimensional file system |
US9025860B2 (en) * | 2012-08-06 | 2015-05-05 | Microsoft Technology Licensing, Llc | Three-dimensional object browsing in documents |
US9772979B1 (en) | 2012-08-08 | 2017-09-26 | Amazon Technologies, Inc. | Reproducing user browsing sessions |
US8943197B1 (en) | 2012-08-16 | 2015-01-27 | Amazon Technologies, Inc. | Automated content update notification |
US9466079B2 (en) * | 2012-09-04 | 2016-10-11 | Cox Enterprises, Inc. | Systems and methods for facilitating the purchase of one or more vehicles |
US11397462B2 (en) * | 2012-09-28 | 2022-07-26 | Sri International | Real-time human-machine collaboration using big data driven augmented reality technologies |
US8793573B2 (en) * | 2012-10-29 | 2014-07-29 | Dropbox, Inc. | Continuous content item view enhanced through smart loading |
US20150286348A1 (en) * | 2012-11-02 | 2015-10-08 | Ge Intelligent Platforms, Inc. | Apparatus and method for intelligent querying and filtering |
US9418145B2 (en) * | 2013-02-04 | 2016-08-16 | TextWise Company, LLC | Method and system for visualizing documents |
US8996511B2 (en) | 2013-03-15 | 2015-03-31 | Envizium, Inc. | System, method, and computer product for providing search results in a hierarchical graphical format |
US9578137B1 (en) | 2013-06-13 | 2017-02-21 | Amazon Technologies, Inc. | System for enhancing script execution performance |
US10152463B1 (en) | 2013-06-13 | 2018-12-11 | Amazon Technologies, Inc. | System for profiling page browsing interactions |
WO2015035172A1 (en) * | 2013-09-05 | 2015-03-12 | White David E | Electronic search engine with enhanced functionalities of generating visual only product/service-relevant search results configured for delivery of product/service-relevant information upon selection thereof, and mobile smart devices and methods of utilizing thereof online |
US9417757B1 (en) * | 2013-09-20 | 2016-08-16 | Emc Corporation | Scalable hierarchical user interface display |
EP3063608B1 (en) | 2013-10-30 | 2020-02-12 | Apple Inc. | Displaying relevant user interface objects |
KR101512084B1 (en) * | 2013-11-15 | 2015-04-17 | 한국과학기술원 | Web search system for providing 3 dimensional web search interface based virtual reality and method thereof |
US20150268825A1 (en) * | 2014-03-18 | 2015-09-24 | Here Global B.V. | Rendering of a media item |
US10613828B2 (en) * | 2014-04-15 | 2020-04-07 | Google Llc | Dynamic and personalized filtering of media content |
US9635041B1 (en) | 2014-06-16 | 2017-04-25 | Amazon Technologies, Inc. | Distributed split browser content inspection and analysis |
KR20170058942A (en) * | 2014-09-22 | 2017-05-29 | 톰슨 라이센싱 | Use of depth perception as indicator of search, user interest or preference |
CN104750816A (en) * | 2015-03-30 | 2015-07-01 | 百度在线网络技术(北京)有限公司 | Information searching method and device |
US10417259B2 (en) * | 2015-05-13 | 2019-09-17 | International Business Machines Corporation | Representing search results via a three-dimensional matrix |
US10824638B2 (en) * | 2016-03-25 | 2020-11-03 | Microsoft Technology Licensing, Llc | Enhancing object representations using inferred user intents |
US10650475B2 (en) * | 2016-05-20 | 2020-05-12 | HomeAway.com, Inc. | Hierarchical panel presentation responsive to incremental search interface |
DK201670595A1 (en) | 2016-06-11 | 2018-01-22 | Apple Inc | Configuring context-specific user interfaces |
US11816325B2 (en) | 2016-06-12 | 2023-11-14 | Apple Inc. | Application shortcuts for carplay |
US10852926B2 (en) * | 2017-05-23 | 2020-12-01 | Salesforce.Com., Inc. | Filter of data presentations via user-generated links |
US11054971B2 (en) | 2017-05-23 | 2021-07-06 | Salesforce.Com., Inc. | Modular runtime environment |
CN107291340B (en) * | 2017-06-12 | 2020-10-09 | 北京小鸟看看科技有限公司 | Method for realizing interface effect, computing equipment and storage medium |
US10726095B1 (en) | 2017-09-26 | 2020-07-28 | Amazon Technologies, Inc. | Network content layout using an intermediary system |
US10664538B1 (en) | 2017-09-26 | 2020-05-26 | Amazon Technologies, Inc. | Data security and data access auditing for network accessible content |
US10523922B2 (en) * | 2018-04-06 | 2019-12-31 | Zspace, Inc. | Identifying replacement 3D images for 2D images via ranking criteria |
US11301050B2 (en) | 2018-06-03 | 2022-04-12 | Apple Inc. | Method and device for presenting a synthesized reality user interface |
US11675476B2 (en) | 2019-05-05 | 2023-06-13 | Apple Inc. | User interfaces for widgets |
CN112486127B (en) * | 2020-12-07 | 2021-12-21 | 北京达美盛软件股份有限公司 | Virtual inspection system of digital factory |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5729673A (en) * | 1995-04-07 | 1998-03-17 | Avid Technology, Inc. | Direct manipulation of two-dimensional moving picture streams in three-dimensional space |
US5911139A (en) * | 1996-03-29 | 1999-06-08 | Virage, Inc. | Visual image database search engine which allows for different schema |
US20030050927A1 (en) * | 2001-09-07 | 2003-03-13 | Araha, Inc. | System and method for location, understanding and assimilation of digital documents through abstract indicia |
US20040225638A1 (en) * | 2003-05-08 | 2004-11-11 | International Business Machines Corporation | Method and system for data mining in high dimensional data spaces |
US20060116994A1 (en) * | 2004-11-30 | 2006-06-01 | Oculus Info Inc. | System and method for interactive multi-dimensional visual representation of information content and properties |
US20070070066A1 (en) * | 2005-09-13 | 2007-03-29 | Bakhash E E | System and method for providing three-dimensional graphical user interface |
US20080155426A1 (en) * | 2006-12-21 | 2008-06-26 | Microsoft Corporation | Visualization and navigation of search results |
US20090112800A1 (en) * | 2007-10-26 | 2009-04-30 | Athellina Rosina Ahmad Athsani | System and method for visual contextual search |
US8498627B2 (en) * | 2011-09-15 | 2013-07-30 | Digimarc Corporation | Intuitive computing methods and systems |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3586549B2 (en) * | 1997-12-08 | 2004-11-10 | 株式会社日立製作所 | Image search method and apparatus |
US6434556B1 (en) * | 1999-04-16 | 2002-08-13 | Board Of Trustees Of The University Of Illinois | Visualization of Internet search information |
US6505194B1 (en) * | 2000-03-29 | 2003-01-07 | Koninklijke Philips Electronics N.V. | Search user interface with enhanced accessibility and ease-of-use features based on visual metaphors |
US6567103B1 (en) * | 2000-08-02 | 2003-05-20 | Verity, Inc. | Graphical search results system and method |
JP2002183150A (en) * | 2000-12-13 | 2002-06-28 | Nec Corp | System and method for presenting position-related information and recording medium with recorded control program thereof |
US20020169759A1 (en) * | 2001-05-14 | 2002-11-14 | International Business Machines Corporation | Method and apparatus for graphically formulating a search query and displaying result set |
US20040216036A1 (en) | 2002-09-13 | 2004-10-28 | Yahoo! Inc. | Browser user interface |
US7117453B2 (en) * | 2003-01-21 | 2006-10-03 | Microsoft Corporation | Media frame object visualization system |
US20040153445A1 (en) * | 2003-02-04 | 2004-08-05 | Horvitz Eric J. | Systems and methods for constructing and using models of memorability in computing and communications applications |
US8555165B2 (en) * | 2003-05-08 | 2013-10-08 | Hillcrest Laboratories, Inc. | Methods and systems for generating a zoomable graphical user interface |
JP4341408B2 (en) * | 2004-01-15 | 2009-10-07 | パナソニック株式会社 | Image display method and apparatus |
US7984389B2 (en) * | 2006-01-28 | 2011-07-19 | Rowan University | Information visualization system |
JP4650293B2 (en) * | 2006-02-15 | 2011-03-16 | 富士フイルム株式会社 | Image classification display device and image classification display program |
KR100898454B1 (en) | 2006-09-27 | 2009-05-21 | 야후! 인크. | Integrated search service system and method |
US7437370B1 (en) * | 2007-02-19 | 2008-10-14 | Quintura, Inc. | Search engine graphical interface using maps and images |
US7941429B2 (en) | 2007-07-10 | 2011-05-10 | Yahoo! Inc. | Interface for visually searching and navigating objects |
WO2009040574A1 (en) * | 2007-09-24 | 2009-04-02 | Taptu Ltd | Search results with search query suggestions |
-
2009
- 2009-08-31 US US12/551,236 patent/US8335784B2/en active Active
-
2010
- 2010-07-26 TW TW099124559A patent/TWI490714B/en not_active IP Right Cessation
- 2010-08-24 KR KR1020127005272A patent/KR101691249B1/en active IP Right Grant
- 2010-08-24 SG SG10201404668VA patent/SG10201404668VA/en unknown
- 2010-08-24 AU AU2010286753A patent/AU2010286753B2/en not_active Ceased
- 2010-08-24 CN CN201080038603.0A patent/CN102483760B/en active Active
- 2010-08-24 EP EP10812535.2A patent/EP2473934A4/en not_active Ceased
- 2010-08-24 WO PCT/US2010/046398 patent/WO2011025749A2/en active Application Filing
- 2010-08-24 BR BR112012004517A patent/BR112012004517A2/en not_active IP Right Cessation
- 2010-08-24 KR KR1020167012787A patent/KR101820256B1/en active IP Right Grant
- 2010-08-24 CA CA2768966A patent/CA2768966C/en not_active Expired - Fee Related
- 2010-08-24 SG SG2012000519A patent/SG177505A1/en unknown
- 2010-08-24 JP JP2012527904A patent/JP5320509B2/en not_active Expired - Fee Related
-
2012
- 2012-11-30 US US13/690,472 patent/US20130091471A1/en not_active Abandoned
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5729673A (en) * | 1995-04-07 | 1998-03-17 | Avid Technology, Inc. | Direct manipulation of two-dimensional moving picture streams in three-dimensional space |
US5911139A (en) * | 1996-03-29 | 1999-06-08 | Virage, Inc. | Visual image database search engine which allows for different schema |
US20030050927A1 (en) * | 2001-09-07 | 2003-03-13 | Araha, Inc. | System and method for location, understanding and assimilation of digital documents through abstract indicia |
US20040225638A1 (en) * | 2003-05-08 | 2004-11-11 | International Business Machines Corporation | Method and system for data mining in high dimensional data spaces |
US20060116994A1 (en) * | 2004-11-30 | 2006-06-01 | Oculus Info Inc. | System and method for interactive multi-dimensional visual representation of information content and properties |
US20070070066A1 (en) * | 2005-09-13 | 2007-03-29 | Bakhash E E | System and method for providing three-dimensional graphical user interface |
US7735018B2 (en) * | 2005-09-13 | 2010-06-08 | Spacetime3D, Inc. | System and method for providing three-dimensional graphical user interface |
US20080155426A1 (en) * | 2006-12-21 | 2008-06-26 | Microsoft Corporation | Visualization and navigation of search results |
US20090112800A1 (en) * | 2007-10-26 | 2009-04-30 | Athellina Rosina Ahmad Athsani | System and method for visual contextual search |
US8498627B2 (en) * | 2011-09-15 | 2013-07-30 | Digimarc Corporation | Intuitive computing methods and systems |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9104621B1 (en) | 2010-09-30 | 2015-08-11 | Axcient, Inc. | Systems and methods for restoring a file |
US9213607B2 (en) | 2010-09-30 | 2015-12-15 | Axcient, Inc. | Systems, methods, and media for synthesizing views of file system backups |
US10284437B2 (en) | 2010-09-30 | 2019-05-07 | Efolder, Inc. | Cloud-based virtual machines and offices |
US9559903B2 (en) | 2010-09-30 | 2017-01-31 | Axcient, Inc. | Cloud-based virtual machines and offices |
US8954544B2 (en) | 2010-09-30 | 2015-02-10 | Axcient, Inc. | Cloud-based virtual machines and offices |
US9235474B1 (en) | 2011-02-17 | 2016-01-12 | Axcient, Inc. | Systems and methods for maintaining a virtual failover volume of a target computing system |
US9785647B1 (en) | 2012-10-02 | 2017-10-10 | Axcient, Inc. | File system virtualization |
US9852140B1 (en) | 2012-11-07 | 2017-12-26 | Axcient, Inc. | Efficient file replication |
US11169714B1 (en) | 2012-11-07 | 2021-11-09 | Efolder, Inc. | Efficient file replication |
US9998344B2 (en) | 2013-03-07 | 2018-06-12 | Efolder, Inc. | Protection status determinations for computing devices |
US10003646B1 (en) | 2013-03-07 | 2018-06-19 | Efolder, Inc. | Protection status determinations for computing devices |
US9397907B1 (en) | 2013-03-07 | 2016-07-19 | Axcient, Inc. | Protection status determinations for computing devices |
US9292153B1 (en) * | 2013-03-07 | 2016-03-22 | Axcient, Inc. | Systems and methods for providing efficient and focused visualization of data |
US9705730B1 (en) | 2013-05-07 | 2017-07-11 | Axcient, Inc. | Cloud storage using Merkle trees |
US10599533B2 (en) | 2013-05-07 | 2020-03-24 | Efolder, Inc. | Cloud storage using merkle trees |
US20220207103A1 (en) * | 2014-02-12 | 2022-06-30 | Pinterest, Inc. | Visual search refinement |
US11714865B2 (en) * | 2014-02-12 | 2023-08-01 | Pinterest, Inc. | Visual search refinement |
US10055092B2 (en) | 2014-05-19 | 2018-08-21 | Samsung Electronics Co., Ltd. | Electronic device and method of displaying object |
US10437423B2 (en) | 2015-10-30 | 2019-10-08 | Hyland Software, Inc. | Methods and apparatuses for providing an infinitely scrolling accumulator |
US20220147531A1 (en) * | 2018-10-29 | 2022-05-12 | State Farm Mutual Automobile Insurance Company | Dynamic data-driven consolidation of user interface interactions requesting roadside assistance |
US11829577B2 (en) * | 2018-10-29 | 2023-11-28 | State Farm Mutual Automobile Insurance Company | Dynamic data-driven consolidation of user interface interactions requesting roadside assistance |
US10838611B1 (en) * | 2019-05-03 | 2020-11-17 | Google Llc | Navigation system on the infinite scroll |
Also Published As
Publication number | Publication date |
---|---|
KR101691249B1 (en) | 2017-01-09 |
BR112012004517A2 (en) | 2016-03-29 |
CN102483760B (en) | 2014-09-24 |
KR20160062182A (en) | 2016-06-01 |
US20110055203A1 (en) | 2011-03-03 |
CA2768966A1 (en) | 2011-03-03 |
AU2010286753B2 (en) | 2015-01-22 |
US8335784B2 (en) | 2012-12-18 |
EP2473934A2 (en) | 2012-07-11 |
AU2010286753A1 (en) | 2012-02-16 |
KR20120073216A (en) | 2012-07-04 |
SG10201404668VA (en) | 2014-10-30 |
KR101820256B1 (en) | 2018-01-18 |
SG177505A1 (en) | 2012-02-28 |
TW201112026A (en) | 2011-04-01 |
CN102483760A (en) | 2012-05-30 |
WO2011025749A2 (en) | 2011-03-03 |
JP5320509B2 (en) | 2013-10-23 |
EP2473934A4 (en) | 2016-10-12 |
WO2011025749A3 (en) | 2011-05-26 |
JP2013503406A (en) | 2013-01-31 |
CA2768966C (en) | 2017-09-26 |
TWI490714B (en) | 2015-07-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8335784B2 (en) | Visual search and three-dimensional results | |
US11907240B2 (en) | Method and system for presenting a search result in a search result card | |
US7941429B2 (en) | Interface for visually searching and navigating objects | |
US6961731B2 (en) | Apparatus and method for organizing and/or presenting data | |
TWI454944B (en) | Systems and methods for providing advanced search result page content | |
US8170916B1 (en) | Related-item tag suggestions | |
TWI585597B (en) | Systems and methods for providing advanced search result page content | |
US8086504B1 (en) | Tag suggestions based on item metadata | |
US9558270B2 (en) | Search result organizing based upon tagging | |
US20130151509A1 (en) | System and method for graphical display of search results | |
US20100306198A1 (en) | System and method for determining categories associated with searches of electronic catalogs and displaying category information with search results | |
US9330071B1 (en) | Tag merging | |
AU2014259978A1 (en) | Tagged search result maintenance | |
US20150170203A1 (en) | Presenting search results | |
US11170062B2 (en) | Structured search via key-objects | |
CN104331430B (en) | The method and device of displaying is patterned to online movie object multiple features dimension | |
US20240095800A1 (en) | Delivery-related search and analytics | |
US10861080B1 (en) | Visualization region search refinement | |
WO2017123746A1 (en) | System and method for intuitive content browsing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:RAY, PAUL RONALD;GUTT, ZACHARY MARK;EDWARDS, RODNEY COLEMAN;AND OTHERS;SIGNING DATES FROM 20090831 TO 20090904;REEL/FRAME:029390/0893 |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034544/0541 Effective date: 20141014 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |