US20060009966A1 - Method and system for extracting information from unstructured text using symbolic machine learning - Google Patents

Method and system for extracting information from unstructured text using symbolic machine learning Download PDF

Info

Publication number
US20060009966A1
US20060009966A1 US10/979,162 US97916204A US2006009966A1 US 20060009966 A1 US20060009966 A1 US 20060009966A1 US 97916204 A US97916204 A US 97916204A US 2006009966 A1 US2006009966 A1 US 2006009966A1
Authority
US
United States
Prior art keywords
pattern
learning
text
generalization
precedence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/979,162
Inventor
David Johnson
Frank Oles
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US10/979,162 priority Critical patent/US20060009966A1/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JOHNSON, DAVID E., OLES, FRANK J.
Publication of US20060009966A1 publication Critical patent/US20060009966A1/en
Priority to US12/507,866 priority patent/US8140323B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing

Definitions

  • the present invention generally relates to extracting information from text. More specifically, in a relational learning system, a pattern learner module receives a small number of learning samples defined by user interactions in relational pattern templates format wherein elements are defined in a precedence relation and in an inclusion relation, and calculates a minimal most specific generalization (MMSG) for these samples so that information matching the generalized template can then be extracted from unseen text.
  • MMSG minimal most specific generalization
  • Extracting relational information from text is an important and unsolved problem in the area of Unstructured Information Management. Many applications including search, question answering, and combining unstructured and structured information could benefit from accurate extraction of relational information.
  • the present invention deals with learning to recognize patterns in text (training data) that characterize the presence of a kind of information in the training data and applying the learned patterns to extract similar kinds of information from new text (unseen text).
  • the present invention deals with learning to recognize patterns that characterize when a particular relation exists between textual elements, mentions of named entities, or phrases that are present in text. This type of learning from text is sometimes called ‘relational learning’.
  • Relation instance Each specific occurrence of a relation is termed a “relation instance”. Often, of particular interest is a relation that describes a kind of event that has occurred, that is occurring or that will occur. In the setting of a relation that describes a kind of event, a relation instance may be termed an “event mention”.
  • a relational learning system might extract the relational information: [Relation: Appoint, Appointee: “Jack Jones”, Role: “CEO of XYZ Corp”].
  • relational learning involves a defined “relation” that includes one or more “parameters” that fit into the relation “template”.
  • “Jack Jones” is the “Appointee”
  • “CEO of XYZ Corp” is the “Role”.
  • “Appointee” and “Role” are the argument names of the “Appoint” relation.
  • the template implies a relationship between the arguments, such as the order of the argument or the interconnection understood by a word or sentence structure, such as a verb or prepositional phrase relationship. The significance of determining this relationship will become apparent as the present invention is further described, since the inventors have recognized that a mere ordering of tokens is insufficient for effective information extraction.
  • a user exemplarily, a non-specialist
  • Machine learning approaches fall into two classes:
  • Machine learning approaches have the advantage that they require only labeled examples of the information sought. Statistical methods are quite popular, but they suffer from the problem of labeling sufficient data accurately for training a model. This is a major problem for such approaches.
  • the defined patterns can then be used to identify relation instances in hitherto unseen sentences with high precision.
  • a method (and structure) of extracting information from text including parsing an input sample of text to form a parse tree and receiving user inputs to define a machine-labeled learning pattern from the parse tree.
  • an apparatus for relational learning including a generator for developing a precedence inclusion (PI) pattern of a learning sample wherein elements in said learning sample are machine-labeled to define a precedence relation and an inclusion relation.
  • PI precedence inclusion
  • a signal-bearing medium tangibly embodying a program of machine-readable instructions executable by a digital processing apparatus to perform a method of relational learning, the machine-readable instructions including a precedence inclusion (PI) pattern learning module for generating a PI pattern of a learning sample wherein elements in said learning sample are machine-labeled to define a precedence relation and an inclusion relation.
  • PI precedence inclusion
  • a method of searching unseen text including at least one of conducting a search of unseen text by developing a precedence inclusion (PI) pattern of at least one learning sample and using the PI pattern for comparison with unseen text and providing a computerized tool to a user for conducting the search.
  • PI precedence inclusion
  • the present invention provides an improved method for relational learning in which a non-specialist can intuitively use the tool that embodies this method to develop a PI pattern template to be used for comparison with unseen text.
  • FIG. 1 illustrates an overview block diagram of an exemplary embodiment 100 of the present invention
  • FIG. 2 illustrates a block diagram of the Linguistic Analyzer 101 ;
  • FIG. 3 shows an example of an actual output display of the Linguistic Analyzer 101 ;
  • FIG. 4 illustrates a block diagram 400 of the PI Pattern Applier 106 ;
  • FIG. 5 illustrates an exemplary screen shot 500 from a demonstration system that illustrates the defining of a relation instance
  • FIG. 6 illustrates an exemplary screen shot 600 from a demonstration system that illustrates the learning of a PI pattern
  • FIG. 7 illustrates an exemplary screen shot 700 from a demonstration system that illustrates the application of a PI pattern
  • FIG. 8 shows an exemplary flowchart 800 of the learning phase
  • FIG. 9 shows an exemplary flowchart 900 of the application phase
  • FIG. 10 illustrates an exemplary block diagram of the major components of the modules of an exemplary software tool 1000 that embodies the present invention
  • FIG. 11 illustrates an exemplary hardware/information handling system 1100 for incorporating the present invention therein.
  • FIG. 12 illustrates a signal bearing medium 1200 (e.g., storage medium) for storing steps of a program of a method according to the present invention.
  • a signal bearing medium 1200 e.g., storage medium
  • FIGS. 1-12 exemplary embodiments of the present invention will now be described.
  • Machine learning approaches have the advantage that they require only labeled examples of the information sought.
  • Much recent work on relational learning has been statistical.
  • One such approach that reflects the state of the art for statistical methods is “Kernel Methods for Relation Extraction” by D. Zelenko, C. Aone, and A. Richardella, where the learning is of a function measuring similarity between shallow parses of examples.
  • Statistical methods in particular, need to have a large amount of labeled training data before anything useful can be done. This is a major problem for statistical approaches.
  • rule-based systems such as the RAPIER system described in “Relational Learning of Pattern-Match Rules for Information Extraction” by M. Califf and R. Mooney, has been attempted.
  • the symbolic approach of the current invention is capable of generalizing from a small number of examples (positive cases) and naturally lends itself to an iterative, interactive learning approach.
  • the current system and method are dynamic.
  • the present invention can use only a few samples for the initial learning.
  • An example is given below in which only two learning samples are used.
  • the tool performs even when only a single learning sample is used, even though a single sample does not fit as neatly into the underlying mathematical theory.
  • the tool is still capable of handling this special case of a single learning sample.
  • the tool heuristically generalizes components for the parameters, such as names, locations, dates, etc., for the single pattern.
  • performance would improve as more learning samples are added, particularly relative to being able to provide a more precisely-defined search template.
  • This learning approach means that a non-expert can use the system and method for learning patterns, simply by providing some examples and indicating which information in the examples should be extracted and what, if anything, the pieces of information should be labeled.
  • the present invention is based on the marriage of deep parsing with a new theory of symbolic pattern generation and, therefore, has a different technical basis from the approaches discussed above. It contrasts sharply with all of the above approaches in that it is based on learning from small sets of sentences that a person would judge to be similar to one another.
  • the approach of the present invention is tailored directly to the vision of the analyst defining the information extraction and the analyst's dynamic needs, in a setting where all the linguistics, computer science, and mathematics are kept under the table (e.g., transparent) so that a non-specialist user can easily and intuitively provide the steps necessary for the preliminary learning phase.
  • the present invention employs the new mathematical theory of precedence-inclusion patterns as the means of learning patterns from a small number of relation instances. These patterns are based on parsed text in which named entity mentions are machine-labeled. The patterns so defined are then applied to other hitherto unseen sentences to find new relation instances with high precision.
  • This new type of symbolic pattern generalization (machine learning) algorithm is used in conjunction with a general purpose parser that produces parse trees of sentences.
  • FIG. 1 An overview of the process 100 is shown in FIG. 1 to demonstrate an exemplary embodiment of the present invention.
  • the process 100 can be broken into two phases, the upper-level phase A i learning process (e.g., A1, A2, A3), and the lower-level phase B i application process (e.g., B1, B2, B3).
  • the upper-level phase A i learning process e.g., A1, A2, A3
  • the lower-level phase B i application process e.g., B1, B2, B3
  • each sample sentence for the preliminary learning process is sequentially used as the basis of defining a machine-labeled relation representing that sample sentence.
  • a user interacts with the linguistic analyzer 101 in a process that includes individually parsing a small number of examples 103 into parse trees 104 , components of which are then provided as inputs into PI pattern learner 102 to generate learning (learned) patterns 105 .
  • the learned patterns 105 then become input patterns into PI Pattern Applier 106 .
  • Input documents from data source 107 are then individually evaluated by the linguistic analyzer 101 , which forms a parse tree 108 of each unseen text document or sentence, which then is compared with the generalization of the learned patterns 105 .
  • the output, the extracted information 109 are unseen text documents that match the pattern of the generalization of the learned patterns 105 .
  • the system also uses a named entity detector to chunk and label names of things, such as people, places, dates, times, monetary amounts, occupations, positions, and organizations, in conjunction with a general parser for the language of interest.
  • the named entity chunking phase reduces the complexity of the parse tree and provides useful class labels for key arguments, such as, in the exemplary scenario, Person, Place, and Organization.
  • Using a general parser means that the syntactic analysis phase of the system does not have to be manually constructed or altered for particular domains or tasks.
  • the invention could be used without a named entity recognition phase, and it could be used with a specialized parser or a partial parser that does not produce complete parses of sentences.
  • the new symbolic pattern generalization algorithm determines, in a mathematically precise sense, the most specific generalization of the set of trees, and this generalization can then be applied to other parse trees (e.g., parse trees 108 ) to determine if any of are instances of the learned generalization.
  • the system them can extract the corresponding information from matched trees.
  • the specific type of structures used in the present invention is original, as is the method of generalization.
  • the structures learned are more general than trees, which are the structures commonly used in computational linguistics.
  • a generalization of a set of trees is also a tree (e.g., a more general tree).
  • the generalization of a set of trees, represented as precedence-inclusion patterns is not, in general, a tree.
  • This generality has practical implications for the invention, as described in the detailed material below, in that a “best generalization” (e.g., a Minimal Most Specific Generalization (MMSG)) is generated for the learned structures.
  • MMSG Minimal Most Specific Generalization
  • other approaches to symbolic generalization are ad hoc in that there is no formal notion of a best generalization.
  • parse trees by themselves, as currently conceived, are, from a technical point of view, seriously deficient as a basis for learning patterns of linguistic constituents of sentences from small numbers of examples.
  • the problem is that, in general, there is no constituent structure tree that functions as a mathematically well-defined “best generalization” of a finite set of constituent structure trees.
  • the present invention is based on the mathematical theory of precedence-inclusion patterns, as discussed in the above-referenced Provisional Application No. 60/586,877, the contents of which are hereby incorporated by reference.
  • FIG. 1 Another non-limiting example of varying the embodiment shown in FIG. 1 includes using heuristic ways to improve the practical utility of the algorithms in an application.
  • synonym sets or online resources such as WordNet, or ontologies, or other techniques from computational linguistics can be used, which, when combined with the fundamental algorithms of the present invention, can result in improved generalization performance.
  • FIG. 2 illustrates a block diagram of components of an exemplary linguistic analyzer 101 used in the present invention.
  • the Named Entity Annotator (NE Annotator) 201 identifies named entities mentioned in text input 103 , 107 and stores the information it finds as text annotations, either as a sequence of characters in text or a sequence of words in text.
  • the parser 202 used in the linguistic analyzer is one that “respects the named entity annotations.” This expression means that the named entities mentions, which may span several words, that are identified in text by the NE annotator 201 are treated as single tokens by the parser 202 .
  • FIG. 3 provides a sample 300 of an actual output 104 , 107 of the Linguistic Analyzer 101 .
  • this output could be due to the action in either the learning phase (e.g., A3) or the application phase (e.g., B3).
  • the input sentence 301 is “Carlson acquired Ask Mr. Foster in 1979.”
  • the named entities 302 from the Named Entity Annotator 201 are also shown. Since named entity annotation preceded parsing, the word “Ask” was included in the name of an organization, and so it was not treated as a verb by the parser, thus improving the accuracy of the parser.
  • FIG. 4 illustrates a block diagram 400 of the PI Pattern Learner 102 .
  • Relation Definition User Interface submodule 401 enables the user to: (1) name a relationship; (2) specify the number of relation arguments and their names; and (3) store this information in memory. It is even possible for a relation to have zero arguments, and this would be the case if the user had an interest in simply defining patterns characteristic of a certain kind of text without identifying specific slot fillers.
  • Relation Instance Definition User Interface submodule 402 enables the user to: (1) choose a relation definition stored in memory by the Relation Definition User Interface 401 ; (2) select a text corpus processed by the Linguistic Analyzer module; (3) choose text, which often, but not necessarily, are single sentences, from the corpus; (4) identify relation instances in the chosen corpus consistent with the chosen relation definition, including none, some, or all, of the slot fillers in the chosen text; and (5) compute, name, and store PI patterns containing this information in memory accompanied by the original text giving rise to the PI pattern.
  • User Interface for the Selection of Instances and for Computing Generalizations submodule 403 enables the user to: (1) view relation instances stored in memory by User Interface submodule 402 ; (2) select two or more of those relation instances; and (3) compute, name, and store in memory a PI pattern that is a generalization of the selected instances.
  • FIGS. 5, 6 , and 7 illustrate exemplary screen displays 500 , 600 , 700 from a prototype demonstration of the tool that embodies methods of the present invention.
  • FIG. 5 shows an exemplary screen display 500 from the demonstration system that illustrates the process of defining a relation instance.
  • the user selects this display 500 by the “Instance Definition” tab 501 in the upper menu bar.
  • the screen 500 contains an upper panel 502 for viewing a corpus of sentences that have undergone linguistic analysis and for selecting a sentence from that corpus by, for example, user inputs from a mouse.
  • the middle panel 503 permits the user to examine a particular sentence that is to be the basis of a pattern containing a relation instance and to select linguistic elements (e.g., a noun phrase, a prepositional phrase, etc.) from that sentence.
  • the bottom panel 504 enables the user to construct a relation instance in a step-by-step fashion based on items selected in the middle panel by associating argument names (e.g., “Previous Position”) with pattern elements (e.g., “vice president and chief financial officer”, which is the head of a noun phrase and which is treated as a single token because it was identified as an OCCUPATION entity in the course of the linguistic analysis of the sentence).
  • argument names e.g., “Previous Position”
  • pattern elements e.g., “vice president and chief financial officer”, which is the head of a noun phrase and which is treated as a single token because it was identified as an OCCUPATION entity in the course of the linguistic analysis of the sentence.
  • the bottom panel 504 also contains a text field 505 in which the relation instance can be given a name and can also be saved, so that it can later be retrieved using that name.
  • FIG. 6 shows an exemplary screen display 600 from the demonstration system that illustrates the learning of a PI pattern. That is, this figure reflects an implementation of submodule 403 (User Interface for the Selection of Instances and for Computing Generalizations submodule).
  • submodule 403 User Interface for the Selection of Instances and for Computing Generalizations submodule.
  • a set of sentences has already been processed in which named entities have been identified and in which the sentences have been parsed.
  • a PI pattern has been generated from the parse by the Relation Instance Definition User Interface 402 and stored in memory.
  • each of these sentences contains an instance of the “ExecutiveChange” relation, which takes five arguments (alternatively termed “slot fillers”) called “Person” 601 , “Previous Position” 602 , “Previous Company” 603 , “New Position” 604 , and “New Company” 605 .
  • the elements of the PI pattern that correspond to slot fillers in these sentences have all been determined by the user, and this information has been stored as part of the stored PI patterns.
  • the MMSG is a smallest representation of exactly the structure common to both the selected instances. It is computed by an implementation of an algorithm given in the theory discussed in the above-referenced Provisional Application and briefly described later.
  • FIG. 7 is a screen shot from a demo system using the pattern learned (e.g., FIG. 6 ) to find hitherto undiscovered instances of the “ExecutiveChange” relation.
  • this figure reflects an implementation of the functionality of the “PI Pattern Applier” Module 106 in FIG. 1 . It is noted that a corpus selection dialog and a means for the user to tell the system where to store the results are not shown in this representation.
  • a set of learned generalizations 105 has been obtained that include the “hasBeenNamed” generalization.
  • the user's intent at this stage is to find new relation instances in this large corpus 107 .
  • FIG. 7 shows that the user has selected the “hasBeenNamed” PI pattern 701 , the “hasBeenNamed” PI pattern has been applied to each sentence in the corpus, and the results of this computation have been stored and displayed to the user.
  • each processed sentence is loaded in turn into memory, (2) then a sentence PI pattern is computed based on the named entities present and the parse of the sentence, and (3) then the system carries out a computation that constructs all pattern-preserving maps from the selected “hasBeenNamed” PI pattern to each sentence PI pattern, provided that such a map exists.
  • the construction of each pattern-preserving map signifies the discovery of an “ExecutiveChange” relation instance, including those slot fillers able to be identified, that can be reported to the user and stored in memory.
  • Two matching sentences 703 , 704 are shown in FIG. 7 .
  • NE Annotation may be based on statistical machine learning, dictionary lookup, hand-written regular expressions, etc.
  • this element is implemented as deep parsing via IBM's Slot Grammar technology, but it can be implemented in any parser that respects the NE Annotation of the first element.
  • the general linguistics analyzer is configurable via ontologies and dictionaries.
  • This new type of symbolic machine learning is based on the new mathematical theory of “Precedence-Inclusion Patterns.”
  • the learning of Precedence-Inclusion patterns is a type of symbolic machine learning based on the new mathematical theory of pattern generalization developed by one of the co-inventors, as presented in the above-referenced Provisional Application.
  • This new theory of symbolic learning is suitable for learning patterns from structures describing how elements are arranged with respect to one another and, in particular, as embodied in the present invention, from parse trees.
  • This technique is capable of learning accurate generalization from a few examples (e.g., only two examples, or even a single example), unlike the current statistical machine learning approaches that require hundreds or thousands of labeled instances of training data.
  • the present invention combines NE annotators, which may be domain-specific, with a general parser to improve the quality of parses.
  • identification of Named Entities will help in learning PI patterns and in applying those patterns to find relation instances in text.
  • Precedence-inclusion patterns are a new class of mathematical objects developed by one of the present inventors that include, but are substantially wider than, constituent structure trees.
  • the theory of PI patterns is explored in detail in the paper “Precedence-Inclusion Patterns and Relational Learning,” by Frank J. Oles, the subject of the above-identified Provisional Application and which paper has been submitted for publication to Theoretical Computer Science.
  • a pattern P generalizes a pattern Q when there is a pattern-preserving map from P to Q.
  • parentheses are placed into sentences to represent sentence structure derivable from parsing, there are pattern-preserving maps from “(Mary went (to the store))” to each of “(Last night Mary went (to the store))”; “(Nary went quickly (to the store))”; and “(Mary went (to the new store))”.
  • PI patterns are sets (1) equipped with two strict partial orders—called precedence and inclusion—that interact with one another through laws called interactive transitivity and interactive irreflexivity, and (2) whose elements may be assigned attributes.
  • PI patterns have a precisely-defined concept of generalization based on the existence of a pattern-preserving mapping from a general pattern to a more specific one, which provides a principled setting for exploring pattern generalization.
  • PI patterns are learned from small numbers of example sentences that contain Event mentions, and the learned PI patterns are then applied to other sentences in order to find new Event mentions.
  • sentences in which named entity mentions are machine-labeled are parsed by ESG.
  • the learning phase 800 proceeds as follows:
  • step 801 the user picks out or creates a similar example sentence describing an Event of interest. It might be preferable, but is not necessary for the operation of the tool, that at least two similar sentences be presented.
  • step 802 the user indicates the Entity mentions that fill the Event's argument roles in each example sentence.
  • step 803 the parser will parse the sentence, and, in steps 804 and 805 the PI Pattern Learner 102 will compute PI patterns for each of the example sentences, including the information about the event of interest and compute the MMSG of the example PI patterns, which will necessarily contain what might be called a “generalized event”.
  • step 806 the user will indicate either that another example sentence will be entered, by choosing the “Instance Definition” tab from the menu at the top of the tool display, thereby returning to step 801 , or will exit the learning phase.
  • the application phase will proceed in the following manner exemplarily shown in the process 900 of FIG. 9 .
  • the user enters this application phase by selecting the “Apply to Text” menu tab 705 (see FIG. 7 ).
  • steps 901 and 902 a sentence is received from the document corpus and parsed into a parse tree.
  • the PI Pattern Applier 106 computes the PI pattern of this sentence and, if the PI pattern is determined to be a special case of the MMSG of the learning phase by computing a pattern-preserving map from the more general pattern to the more specific one if there is one, determines the fillers for the argument roles of the event mentions in the sentence. If a pattern-preserving mapping was found in the previous step, in step 905 the PI Pattern Applier 106 determine the fillers for the argument roles of the Event mentions(s) in the sentence. In step 906 , this sequence is continued for all sentences of the corpus.
  • FIG. 10 illustrates an exemplary block diagram 1000 of the major components that might comprise a computer tool for implementing the present invention.
  • Graphical User Interface (GUI) module 1001 allows the user to provide inputs 1002 and provides the display information 1003 for the screen displays, such as discussed for FIGS. 5-7 .
  • Control module 1004 provides the control instructions to interconnect the various modules and to control the information passing therebetween.
  • the functions of the linguistic analyzer 101 , PI pattern learner 102 , and PI pattern applier 106 have been previously discussed.
  • Memory interface 1005 controls the flow of information between local memory, as well as the flow of information to a possible database such as might contain the corpus documents for either the learning examples 103 or the input documents 107 used in the search phase.
  • the tool 1000 would include a network interface 1007 , exemplarily shown in FIG. 10 as a submodule of the control module 1004 .
  • the present invention has been discussed as used to execute a search through a database or a series of databases or data sources, it should not be considered as being so limited.
  • the present invention might be utilized in environments in which there is no fixed database or data sources but, rather, a real-time data source.
  • the present invention might be used for searching real-time text data in an intelligence-gathering environment, wherein no discrete and readily-identifiable database exists.
  • This corpus containing 280 sentences and headlines, was constructed from short news stories found on the ExecutiveSelect.com website. These stories dealt with executive position changes. Named entity mentions, specifically PERSON, LOCATION, OCCUPATION, and ORGANIZATION, were marked in the corpus by hand, and all sentences were parsed in the corpus with ESG.
  • ACE Automatic Content Extraction
  • each ExecutiveChange Event could be viewed as a template with five slots to fill.
  • first example sentence see item 606 of FIG. 6 :
  • the ExecutiveChange Event in the second sentence is similarly identified, and can be displayed as follows:
  • MMSG of these two patterns which is not a parse tree, and in which there is a generalized ExecutiveChange relation instance that says nothing specific about the text that appears in the slot fillers.
  • the MMSG of these 2 patterns has 15 elements.
  • the constructed generalized ExecutiveChange relation instance is partially described as:
  • the MMSG has more elements than the smaller of the two example patterns that generated it. This is an indication of the fact that one does not get an MMSG by deleting pattern elements in one PI pattern that somehow do not match any element in another PI pattern.
  • thirteen new mentions of the ExecutiveChange Event will be discovered in a few seconds, such as:
  • the new ExecutiveChange relation instance that was found is:
  • the mathematical theory encompasses more than the method discussed above to extract information from unstructured text, since the theory additionally applies to extraction of information from images.
  • the present invention can be viewed as a concrete implementation of a somewhat narrow, specific application of this new mathematical theory as it relates to text. That is, the tool of the present invention is not currently implemented to apply the precedence inclusion pattern theory to image analysis, and it is expected that several problems need to be solved before the theory can be embodied in a tool for images. However, the interested reader is invited to read this paper for more details of this mathematical theory than is discussed below.
  • a pattern extracted from a video may have as its elements some picture elements occurring in individual frames of the video.
  • parsing text e.g., the subject of the present invention
  • linguistic entities in a parse tree may precede one another or they may contain one another.
  • categories of what is referred to in the paper as “2-patterns” include among their objects constituent structure trees, as they are normally defined in computational linguistics.
  • Interactive transitivity is an extension of the concept of a transitive, binary relation to an ordered pair of binary relations.
  • the ordered pair of relations [is to the left of, contains] is interactively transitive, as also would be the ordered pair [is above, contains].
  • a precedence-inclusion pattern is a set equipped with a strictly partially ordered set of strict partial orders, along with some additional structure, in which the strict partial order on the strict partial orders is taken to assert that each related pair of strict partial orders obeys the axioms of interactive transitivity.
  • a precise definition starts in an algebraic style.
  • (O, A, L) be a pattern signature. It is said that a set is a ⁇ -pattern when every ⁇ O has an interpretation ⁇ ,P as a strict partial order on P, along with a partial function ⁇ P : P ⁇ L, called the argument namingfunction, and a total function ⁇ P : P ⁇ L, called the labeling function, such that ⁇ implies that the ordered pair of relations [ ⁇ ,P , ⁇ ,P ] is interactively transitive.
  • ⁇ -pattern When ⁇ is clear from context, a ⁇ -pattern is called a precedence-inclusion pattern. Thus, when the order symbol set O is empty, ⁇ -patterns are just sets with some additional structure. When the order symbol set is one-element set, then ⁇ -patterns are strictly partially ordered sets with some additional structure.
  • a ⁇ -pattern Q is a generalization of a ⁇ -pattern if there is a pattern-preserving map, i.e., a morphism in the category of ⁇ -patterns, from Q to P.
  • msg most specific generalization
  • lgg least general generalization
  • a minimal most specific generalization of a set P of patterns is an msg of P no subpattern of which is an msg of P. These are the kinds of generalizations that are desired.
  • a retraction of a precedence-inclusion pattern P is an idempotent endomorphism r: P ⁇ P, and the set of fixed points of a retraction defines a pattern called a retract of P. A pattern having no proper retracts is said to be fully retracted.
  • FIG. 11 illustrates a typical hardware configuration of an information handling/computer system in accordance with the invention and which preferably has at least one processor or central processing unit (CPU) 1111 .
  • processor or central processing unit
  • the CPUs 1111 are interconnected via a system bus 1112 to a random access memory (RAM) 1114 , read-only memory (ROM) 1116 , input/output (I/O) adapter 1118 (for connecting peripheral devices such as disk units 1121 and tape drives 1140 to the bus 1112 ), user interface adapter 1122 (for connecting a keyboard 1124 , mouse 1126 , speaker 1128 , microphone 1132 , and/or other user interface device to the bus 1112 ), a communication adapter 1134 for connecting an information handling system to a data processing network, the Internet, an Intranet, a personal area network (PAN), etc., and a display adapter 1136 for connecting the bus 1112 to a display device 1138 and/or printer 1139 (e.g., a digital printer or the like).
  • RAM random access memory
  • ROM read-only memory
  • I/O input/output
  • I/O input/output
  • user interface adapter 1122 for connecting a keyboard 1124 , mouse 1126
  • a different aspect of the invention includes a computer-implemented method for performing the above method. As an example, this method may be implemented in the particular environment discussed above.
  • Such a method may be implemented, for example, by operating a computer, as embodied by a digital data processing apparatus, to execute a sequence of machine-readable instructions. These instructions may reside in various types of signal-bearing media.
  • this aspect of the present invention is directed to a programmed product, comprising signal-bearing media tangibly embodying a program of machine-readable instructions executable by a digital data processor incorporating the CPU 1111 and hardware above, to perform the method of the invention.
  • This signal-bearing media may include, for example, a RAM contained within the CPU 1111 , as represented by the fast-access storage for example.
  • the instructions may be contained in another signal-bearing media, such as a magnetic data storage diskette 1200 ( FIG. 12 ), directly or indirectly accessible by the CPU 1111 .
  • the instructions may be stored on a variety of machine-readable data storage media, such as DASD storage (e.g., a conventional “hard drive” or a RAID array), magnetic tape, electronic read-only memory (e.g., ROM, EPROM, or EEPROM), an optical storage device (e.g. CD-ROM, WORM, DVD, digital optical tape, etc.), paper “punch” cards, or other suitable signal-bearing media including transmission media such as digital and analog and communication links and wireless.
  • DASD storage e.g., a conventional “hard drive” or a RAID array
  • magnetic tape e.g., magnetic tape, electronic read-only memory (e.g., ROM, EPROM, or EEPROM), an optical storage device (e.g. CD-ROM, WORM, DVD, digital optical tape, etc.), paper “punch” cards, or other suitable signal-bearing media including transmission media such as digital and analog and communication links and wireless.
  • the machine-readable instructions may comprise software object code.
  • the present invention can be used as a basis for a service or business method.
  • the present invention might be, for example, implemented as a search method for specific databases or, more generally, for any number of data sources available through a computer network, such as the Internet.
  • the present invention is intended as covering, not only the tool itself that executes the methods described above, but also a service made available for using this method by others.
  • an entity might provide the service of conducting searches of databases or the Internet, using the method of the present invention.
  • the present invention is intended as including this exploitation of executing the above-described methods by providing a service to others to conduct searches using these methods.
  • Another service related to the present invention might be based on making a tool available to others so that these clients or customers can themselves use the tool to conduct their own search.
  • a number of non-limiting examples for which it might be employed might include: a web search, wherein a user picks examples and searches for the same relations; intelligence gathering by intelligence agencies; searches by financial analysts, looking for facts, press releases, or SEC filings; or extracting information from clinical notes of medical practitioners.
  • the searches might be done on propriety or publicly available databases, a computer network, including the Internet, or might even be done in real-time.

Abstract

A method (and structure) of extracting information from text, includes parsing an input sample of text to form a parse tree and using user inputs to define a machine-labeled learning pattern from the parse tree.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present Application is related to U.S. Provisional Patent Application No. 60/586,877, filed on Jul. 12, 2004, to Johnson et al., entitled “System and Method for Extracting Information from Unstructured Text Using Symbolic Machine Learning”, having IBM Docket YOR920040239US1, assigned to the present assignee, and incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention generally relates to extracting information from text. More specifically, in a relational learning system, a pattern learner module receives a small number of learning samples defined by user interactions in relational pattern templates format wherein elements are defined in a precedence relation and in an inclusion relation, and calculates a minimal most specific generalization (MMSG) for these samples so that information matching the generalized template can then be extracted from unseen text.
  • 2. Description of the Related Art
  • Extracting relational information from text is an important and unsolved problem in the area of Unstructured Information Management. Many applications including search, question answering, and combining unstructured and structured information could benefit from accurate extraction of relational information.
  • The present invention deals with learning to recognize patterns in text (training data) that characterize the presence of a kind of information in the training data and applying the learned patterns to extract similar kinds of information from new text (unseen text). In particular, the present invention deals with learning to recognize patterns that characterize when a particular relation exists between textual elements, mentions of named entities, or phrases that are present in text. This type of learning from text is sometimes called ‘relational learning’.
  • Each specific occurrence of a relation is termed a “relation instance”. Often, of particular interest is a relation that describes a kind of event that has occurred, that is occurring or that will occur. In the setting of a relation that describes a kind of event, a relation instance may be termed an “event mention”.
  • For instance, from a sentence such as “Jack Jones was appointed CEO of XYZ Corp last week”, a relational learning system might extract the relational information: [Relation: Appoint, Appointee: “Jack Jones”, Role: “CEO of XYZ Corp”].
  • From this example, it can be understood that relational learning involves a defined “relation” that includes one or more “parameters” that fit into the relation “template”. In the example, “Jack Jones” is the “Appointee” and “CEO of XYZ Corp” is the “Role”. “Appointee” and “Role” are the argument names of the “Appoint” relation. It should also be noted that the template implies a relationship between the arguments, such as the order of the argument or the interconnection understood by a word or sentence structure, such as a verb or prepositional phrase relationship. The significance of determining this relationship will become apparent as the present invention is further described, since the inventors have recognized that a mere ordering of tokens is insufficient for effective information extraction.
  • As an exemplary scenario for which the present invention might be used is one in which a user (exemplarily, a non-specialist) wishes to search a database or perhaps the Internet to find data items that, for example, identify CEOs of corporations.
  • Basically, there are currently two main approaches for this relational learning problem:
      • (1) manual development of patterns; and
      • (2) learning patterns using machine learning techniques.
  • Manual approaches are very costly to develop, since they require experts in computational linguistics or related disciplines to develop formal grammars or special purpose programs. Non-specialists cannot customize manual systems for new domains, tasks or languages.
  • Machine learning approaches fall into two classes:
      • (i) statistical approaches; and
      • (ii) symbolic approaches.
  • Machine learning approaches have the advantage that they require only labeled examples of the information sought. Statistical methods are quite popular, but they suffer from the problem of labeling sufficient data accurately for training a model. This is a major problem for such approaches.
  • Moreover, as the relations of interest vary from task to task and even from individual to individual, methods are needed to learn how to extract relations of interest on demand. Further, it would be desirable that non-specialists be able to use the relational learning tool.
  • There are currently no adequate solutions to the problem of trainable relation extraction systems, especially no adequate systems that can be used by non-specialists.
  • Thus, a need continues for a method and system that, as relations of interest vary from task to task and even from individual to individual, learn how to extract relations of interest on demand. Further, it would be desirable that non-specialists be easily able to use a relational learning system.
  • SUMMARY OF THE INVENTION
  • In view of the foregoing, and other, exemplary problems, drawbacks, and disadvantages of the conventional system, it is an exemplary feature of the present invention to provide a system and method for developing learning patterns that can then be used to automatically extract relations from text.
  • It is another exemplary feature of the present invention to provide a technique that can be used by non-specialists.
  • It is another exemplary feature of the present invention to provide a method that allows a user with no special knowledge of linguistics to dynamically define patterns on the basis of a small number of example sentences or pseudo-examples in which the user has marked those named entity mentions that are involved in a relation instance. The defined patterns can then be used to identify relation instances in hitherto unseen sentences with high precision.
  • It is another exemplary feature of the present invention to provide a method in which only a few samples are needed to define a relation pattern for use in searching for similar patterns, including the case in which a single learning sample can be used.
  • To achieve the above exemplary features and others, in a first exemplary aspect of the present invention, described herein is a method (and structure) of extracting information from text, including parsing an input sample of text to form a parse tree and receiving user inputs to define a machine-labeled learning pattern from the parse tree.
  • In a second exemplary aspect of the present invention, described herein is an apparatus for relational learning, including a generator for developing a precedence inclusion (PI) pattern of a learning sample wherein elements in said learning sample are machine-labeled to define a precedence relation and an inclusion relation.
  • In a third exemplary aspect of the present invention, described herein is a signal-bearing medium tangibly embodying a program of machine-readable instructions executable by a digital processing apparatus to perform a method of relational learning, the machine-readable instructions including a precedence inclusion (PI) pattern learning module for generating a PI pattern of a learning sample wherein elements in said learning sample are machine-labeled to define a precedence relation and an inclusion relation.
  • In a fourth exemplary aspect of the present invention, also described herein is a method of searching unseen text, the method including at least one of conducting a search of unseen text by developing a precedence inclusion (PI) pattern of at least one learning sample and using the PI pattern for comparison with unseen text and providing a computerized tool to a user for conducting the search.
  • Thus, the present invention provides an improved method for relational learning in which a non-specialist can intuitively use the tool that embodies this method to develop a PI pattern template to be used for comparison with unseen text.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The foregoing and other exemplary features, aspects and advantages will be better understood from the following detailed description of an exemplary embodiment of the invention with reference to the drawings, in which:
  • FIG. 1 illustrates an overview block diagram of an exemplary embodiment 100 of the present invention;
  • FIG. 2 illustrates a block diagram of the Linguistic Analyzer 101;
  • FIG. 3 shows an example of an actual output display of the Linguistic Analyzer 101;
  • FIG. 4 illustrates a block diagram 400 of the PI Pattern Applier 106;
  • FIG. 5 illustrates an exemplary screen shot 500 from a demonstration system that illustrates the defining of a relation instance;
  • FIG. 6 illustrates an exemplary screen shot 600 from a demonstration system that illustrates the learning of a PI pattern;
  • FIG. 7 illustrates an exemplary screen shot 700 from a demonstration system that illustrates the application of a PI pattern;
  • FIG. 8 shows an exemplary flowchart 800 of the learning phase;
  • FIG. 9 shows an exemplary flowchart 900 of the application phase;
  • FIG. 10 illustrates an exemplary block diagram of the major components of the modules of an exemplary software tool 1000 that embodies the present invention;
  • FIG. 11 illustrates an exemplary hardware/information handling system 1100 for incorporating the present invention therein; and
  • FIG. 12 illustrates a signal bearing medium 1200 (e.g., storage medium) for storing steps of a program of a method according to the present invention.
  • DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS OF THE INVENTION
  • Referring now to the drawings, and more particularly to FIGS. 1-12, exemplary embodiments of the present invention will now be described.
  • Machine learning approaches have the advantage that they require only labeled examples of the information sought. Much recent work on relational learning has been statistical. One such approach that reflects the state of the art for statistical methods is “Kernel Methods for Relation Extraction” by D. Zelenko, C. Aone, and A. Richardella, where the learning is of a function measuring similarity between shallow parses of examples. Statistical methods, in particular, need to have a large amount of labeled training data before anything useful can be done. This is a major problem for statistical approaches.
  • Work in another vein has concerned various attempts to accomplish relational learning by using heuristics to learn finite state recognizers or regular expressions, as exemplified by “Learning Information Extraction Rules for Semi-Structure and Free Text”, by S. Soderland.
  • Finally, the automatic construction of rule-based systems, such as the RAPIER system described in “Relational Learning of Pattern-Match Rules for Information Extraction” by M. Califf and R. Mooney, has been attempted.
  • In contrast, as described in the following discussion concerning the overview in FIG. 1 of an exemplary embodiment, the symbolic approach of the current invention is capable of generalizing from a small number of examples (positive cases) and naturally lends itself to an iterative, interactive learning approach.
  • That is, in an interactive approach, a user might use the tool of the present invention to select a few data as learning samples, use the tool of the present invention to generalize these learning samples, test its performance, and, if desirable, pick additional samples to improve the learning and thereby improve performance. Thus, the current system and method are dynamic.
  • As mentioned, in contrast to conventional methods, the present invention can use only a few samples for the initial learning. An example is given below in which only two learning samples are used. It is noted that the tool performs even when only a single learning sample is used, even though a single sample does not fit as neatly into the underlying mathematical theory. However, the tool is still capable of handling this special case of a single learning sample. In this special case, the tool heuristically generalizes components for the parameters, such as names, locations, dates, etc., for the single pattern. However, in general, it would reasonably be expected that performance would improve as more learning samples are added, particularly relative to being able to provide a more precisely-defined search template.
  • This learning approach, together with the other aspects of the invention, means that a non-expert can use the system and method for learning patterns, simply by providing some examples and indicating which information in the examples should be extracted and what, if anything, the pieces of information should be labeled.
  • It can be said that the present invention is based on the marriage of deep parsing with a new theory of symbolic pattern generation and, therefore, has a different technical basis from the approaches discussed above. It contrasts sharply with all of the above approaches in that it is based on learning from small sets of sentences that a person would judge to be similar to one another.
  • Moreover, it will readily recognized that the approach of the present invention is tailored directly to the vision of the analyst defining the information extraction and the analyst's dynamic needs, in a setting where all the linguistics, computer science, and mathematics are kept under the table (e.g., transparent) so that a non-specialist user can easily and intuitively provide the steps necessary for the preliminary learning phase.
  • The present invention employs the new mathematical theory of precedence-inclusion patterns as the means of learning patterns from a small number of relation instances. These patterns are based on parsed text in which named entity mentions are machine-labeled. The patterns so defined are then applied to other hitherto unseen sentences to find new relation instances with high precision. This new type of symbolic pattern generalization (machine learning) algorithm is used in conjunction with a general purpose parser that produces parse trees of sentences.
  • An overview of the process 100 is shown in FIG. 1 to demonstrate an exemplary embodiment of the present invention. The process 100 can be broken into two phases, the upper-level phase Ai learning process (e.g., A1, A2, A3), and the lower-level phase Bi application process (e.g., B1, B2, B3).
  • In the first phase A, each sample sentence for the preliminary learning process is sequentially used as the basis of defining a machine-labeled relation representing that sample sentence. In this step, a user interacts with the linguistic analyzer 101 in a process that includes individually parsing a small number of examples 103 into parse trees 104, components of which are then provided as inputs into PI pattern learner 102 to generate learning (learned) patterns 105.
  • In the second phase B, the learned patterns 105 then become input patterns into PI Pattern Applier 106. Input documents from data source 107 are then individually evaluated by the linguistic analyzer 101, which forms a parse tree 108 of each unseen text document or sentence, which then is compared with the generalization of the learned patterns 105. The output, the extracted information 109, are unseen text documents that match the pattern of the generalization of the learned patterns 105.
  • In an exemplary embodiment shown in FIG. 1, the system also uses a named entity detector to chunk and label names of things, such as people, places, dates, times, monetary amounts, occupations, positions, and organizations, in conjunction with a general parser for the language of interest. The named entity chunking phase reduces the complexity of the parse tree and provides useful class labels for key arguments, such as, in the exemplary scenario, Person, Place, and Organization.
  • Using a general parser means that the syntactic analysis phase of the system does not have to be manually constructed or altered for particular domains or tasks. However, in principle, the invention could be used without a named entity recognition phase, and it could be used with a specialized parser or a partial parser that does not produce complete parses of sentences.
  • More specifically, in the present invention, given a set of parse trees of, presumably, related sentences (e.g., the learning samples 103), the new symbolic pattern generalization algorithm determines, in a mathematically precise sense, the most specific generalization of the set of trees, and this generalization can then be applied to other parse trees (e.g., parse trees 108) to determine if any of are instances of the learned generalization.
  • In addition, by associating pieces of information in the example sentences with elements in a relation or template, the system them can extract the corresponding information from matched trees. The specific type of structures used in the present invention is original, as is the method of generalization. Specifically, the structures learned (so-called “precedence inclusion patterns”) are more general than trees, which are the structures commonly used in computational linguistics.
  • In the state of the art, a generalization of a set of trees is also a tree (e.g., a more general tree). In contrast, within the theory applied here, the generalization of a set of trees, represented as precedence-inclusion patterns, is not, in general, a tree. This generality has practical implications for the invention, as described in the detailed material below, in that a “best generalization” (e.g., a Minimal Most Specific Generalization (MMSG)) is generated for the learned structures. In contrast, other approaches to symbolic generalization are ad hoc in that there is no formal notion of a best generalization.
  • That is, the present inventors consider that parse trees by themselves, as currently conceived, are, from a technical point of view, seriously deficient as a basis for learning patterns of linguistic constituents of sentences from small numbers of examples. The problem is that, in general, there is no constituent structure tree that functions as a mathematically well-defined “best generalization” of a finite set of constituent structure trees.
  • To overcome this deficiency, the present invention is based on the mathematical theory of precedence-inclusion patterns, as discussed in the above-referenced Provisional Application No. 60/586,877, the contents of which are hereby incorporated by reference.
  • As stated above, other known approaches to symbolic generalization are ad hoc, in that there is no formal notion of a “best generalization.” In contrast, the approach of the present invention is very general, lending itself to learning over many types of structures: full parse trees, partial parses. Indeed, patterns can be generalized across sentences. The approach has been implemented and another key part of the invention deals with the efficient implementation of the generalization process.
  • Another non-limiting example of varying the embodiment shown in FIG. 1 includes using heuristic ways to improve the practical utility of the algorithms in an application. As non-limiting examples, synonym sets or online resources, such as WordNet, or ontologies, or other techniques from computational linguistics can be used, which, when combined with the fundamental algorithms of the present invention, can result in improved generalization performance.
  • FIG. 2 illustrates a block diagram of components of an exemplary linguistic analyzer 101 used in the present invention. The Named Entity Annotator (NE Annotator) 201 identifies named entities mentioned in text input 103, 107 and stores the information it finds as text annotations, either as a sequence of characters in text or a sequence of words in text.
  • The parser 202 used in the linguistic analyzer is one that “respects the named entity annotations.” This expression means that the named entities mentions, which may span several words, that are identified in text by the NE annotator 201 are treated as single tokens by the parser 202.
  • FIG. 3 provides a sample 300 of an actual output 104, 107 of the Linguistic Analyzer 101. As shown in FIG. 1, this output could be due to the action in either the learning phase (e.g., A3) or the application phase (e.g., B3). The input sentence 301 is “Carlson acquired Ask Mr. Foster in 1979.” The named entities 302 from the Named Entity Annotator 201 are also shown. Since named entity annotation preceded parsing, the word “Ask” was included in the name of an organization, and so it was not treated as a verb by the parser, thus improving the accuracy of the parser.
  • FIG. 4 illustrates a block diagram 400 of the PI Pattern Learner 102. Relation Definition User Interface submodule 401 enables the user to: (1) name a relationship; (2) specify the number of relation arguments and their names; and (3) store this information in memory. It is even possible for a relation to have zero arguments, and this would be the case if the user had an interest in simply defining patterns characteristic of a certain kind of text without identifying specific slot fillers.
  • Relation Instance Definition User Interface submodule 402 enables the user to: (1) choose a relation definition stored in memory by the Relation Definition User Interface 401; (2) select a text corpus processed by the Linguistic Analyzer module; (3) choose text, which often, but not necessarily, are single sentences, from the corpus; (4) identify relation instances in the chosen corpus consistent with the chosen relation definition, including none, some, or all, of the slot fillers in the chosen text; and (5) compute, name, and store PI patterns containing this information in memory accompanied by the original text giving rise to the PI pattern.
  • User Interface for the Selection of Instances and for Computing Generalizations submodule 403 enables the user to: (1) view relation instances stored in memory by User Interface submodule 402; (2) select two or more of those relation instances; and (3) compute, name, and store in memory a PI pattern that is a generalization of the selected instances.
  • It is noted that, generally, it is neither practical nor useful to view the complete structure of relation instances, due to their complexity. Instead, just critical parts, such as the original text, the name of the relation, and the text filling the slots (e.g., the relation arguments), normally need to be viewed.
  • FIGS. 5, 6, and 7 illustrate exemplary screen displays 500, 600, 700 from a prototype demonstration of the tool that embodies methods of the present invention.
  • FIG. 5 shows an exemplary screen display 500 from the demonstration system that illustrates the process of defining a relation instance. The user selects this display 500 by the “Instance Definition” tab 501 in the upper menu bar. The screen 500 contains an upper panel 502 for viewing a corpus of sentences that have undergone linguistic analysis and for selecting a sentence from that corpus by, for example, user inputs from a mouse.
  • The middle panel 503 permits the user to examine a particular sentence that is to be the basis of a pattern containing a relation instance and to select linguistic elements (e.g., a noun phrase, a prepositional phrase, etc.) from that sentence. The bottom panel 504 enables the user to construct a relation instance in a step-by-step fashion based on items selected in the middle panel by associating argument names (e.g., “Previous Position”) with pattern elements (e.g., “vice president and chief financial officer”, which is the head of a noun phrase and which is treated as a single token because it was identified as an OCCUPATION entity in the course of the linguistic analysis of the sentence).
  • Note that the bottom panel 504 also contains a text field 505 in which the relation instance can be given a name and can also be saved, so that it can later be retrieved using that name.
  • FIG. 6 shows an exemplary screen display 600 from the demonstration system that illustrates the learning of a PI pattern. That is, this figure reflects an implementation of submodule 403 (User Interface for the Selection of Instances and for Computing Generalizations submodule).
  • At the start of this stage, a set of sentences has already been processed in which named entities have been identified and in which the sentences have been parsed. For each sentence, based on user inputs, a PI pattern has been generated from the parse by the Relation Instance Definition User Interface 402 and stored in memory.
  • Using the screen 500 exemplarily illustrated in FIG. 5, a user has determined that each of these sentences contains an instance of the “ExecutiveChange” relation, which takes five arguments (alternatively termed “slot fillers”) called “Person” 601, “Previous Position” 602, “Previous Company” 603, “New Position” 604, and “New Company” 605. The elements of the PI pattern that correspond to slot fillers in these sentences have all been determined by the user, and this information has been stored as part of the stored PI patterns.
  • Two of the sentences 606, 607 have been selected and displayed in FIG. 6. The basis for selection was the user's judgment that the sentences were structurally similar. The user chose and entered the name “hasBeenNamed” for the generalization 608, which caused the system to compute and store a Minimal Most Specific Generalization (MMSG) 609 of the two selections.
  • The MMSG is a smallest representation of exactly the structure common to both the selected instances. It is computed by an implementation of an algorithm given in the theory discussed in the above-referenced Provisional Application and briefly described later.
  • FIG. 7 is a screen shot from a demo system using the pattern learned (e.g., FIG. 6) to find hitherto undiscovered instances of the “ExecutiveChange” relation. Thus, this figure reflects an implementation of the functionality of the “PI Pattern Applier” Module 106 in FIG. 1. It is noted that a corpus selection dialog and a means for the user to tell the system where to store the results are not shown in this representation.
  • In relating FIG. 7 back to the view of FIG. 1, at the start of this stage represented in FIG. 7, a set of learned generalizations 105 has been obtained that include the “hasBeenNamed” generalization. There is also a corpus of sentences 107 processed by the Linguistic Analyzer 101, in which named entities have been identified, and in which the sentences have been parsed (e.g., 108). The user's intent at this stage (e.g., 109) is to find new relation instances in this large corpus 107.
  • FIG. 7 shows that the user has selected the “hasBeenNamed” PI pattern 701, the “hasBeenNamed” PI pattern has been applied to each sentence in the corpus, and the results of this computation have been stored and displayed to the user.
  • When the “hasBeenNamed” pattern is applied to the corpus (e.g, by using selector 702), (1) each processed sentence is loaded in turn into memory, (2) then a sentence PI pattern is computed based on the named entities present and the parse of the sentence, and (3) then the system carries out a computation that constructs all pattern-preserving maps from the selected “hasBeenNamed” PI pattern to each sentence PI pattern, provided that such a map exists. The construction of each pattern-preserving map signifies the discovery of an “ExecutiveChange” relation instance, including those slot fillers able to be identified, that can be reported to the user and stored in memory. Two matching sentences 703, 704 are shown in FIG. 7.
  • It is also noted the this stage shown in FIG. 7 was arrived at by selecting the “Apply to Text” tab 705 at the top of the Graphical User Interface display. Similarly, the stage shown in FIG. 6 was initiated by selecting the “Select & Learn” tab 610.
  • From the above description, it can now be recognized that the exemplary relation extraction system of the present invention can be described as combining the following three elements:
  • 1. Named Entity (NE) Annotation
  • This terminology is also called “Named Entity Mention Detection”. NE Annotation may be based on statistical machine learning, dictionary lookup, hand-written regular expressions, etc.
  • 2. General Linguistic Analysis
  • In an exemplary embodiment, this element is implemented as deep parsing via IBM's Slot Grammar technology, but it can be implemented in any parser that respects the NE Annotation of the first element. The general linguistics analyzer is configurable via ontologies and dictionaries.
  • 3. New Approach to Relation Extraction
  • This new type of symbolic machine learning is based on the new mathematical theory of “Precedence-Inclusion Patterns.” The learning of Precedence-Inclusion patterns is a type of symbolic machine learning based on the new mathematical theory of pattern generalization developed by one of the co-inventors, as presented in the above-referenced Provisional Application. This new theory of symbolic learning is suitable for learning patterns from structures describing how elements are arranged with respect to one another and, in particular, as embodied in the present invention, from parse trees. This technique is capable of learning accurate generalization from a few examples (e.g., only two examples, or even a single example), unlike the current statistical machine learning approaches that require hundreds or thousands of labeled instances of training data.
  • An advantage of using the general linguistics analysis is that the clause-level structure of language becomes domain-independent.
  • However, general linguistic analysis has a disadvantage in that general parsers are typically less accurate than domain-specific parsers. Additionally, special expertise is typically required to extend or correct general parsers.
  • To overcome these problems, the present invention combines NE annotators, which may be domain-specific, with a general parser to improve the quality of parses. As an added benefit, identification of Named Entities will help in learning PI patterns and in applying those patterns to find relation instances in text.
  • Precedence-inclusion patterns, “PI patterns” for short, are a new class of mathematical objects developed by one of the present inventors that include, but are substantially wider than, constituent structure trees. The theory of PI patterns is explored in detail in the paper “Precedence-Inclusion Patterns and Relational Learning,” by Frank J. Oles, the subject of the above-identified Provisional Application and which paper has been submitted for publication to Theoretical Computer Science.
  • To further explain the concepts present in the Theory of Precedence-Inclusion Patterns, the following sentence “Mary went to the store” can be said to have a pattern. The same pattern appears in the following three sentences: “Last night, Mary went to the store”; “Mary went quickly to the store”; and “Mary went to the new store.”
  • However, the basic pattern is not present in “Mary went to the movies after the store closed,” even though this last sentence contains the same sequence of tokens: “Mary went”, “to”, “the store.”
  • The conclusion to be reached is that patterns in text should involve more than identifying sequences of tokens, even tokens with types assigned. Therefore, relative to the present invention, in defining text-based patterns, two interrelated strict partial orders are used:
      • 1. Strictly precedes: x<y, as illustrated by:
        • Mary<went<to the store.
      • 2. Strictly includes: x⊃y, as illustrated by:
        • to the store⊃store.
  • It should be noted that the above sentences demonstrate that the partial order concept is related to the analysis of language.
  • A pattern P generalizes a pattern Q when there is a pattern-preserving map from P to Q. Thus, if parentheses are placed into sentences to represent sentence structure derivable from parsing, there are pattern-preserving maps from “(Mary went (to the store))” to each of “(Last night Mary went (to the store))”; “(Nary went quickly (to the store))”; and “(Mary went (to the new store))”. But, there is no pattern preserving map from “(Mary went to the store)” to “(Mary went (to the movies) (after the store closed))”, since the parenthetic structures derived from parsing are inconsistent in these two sentences.
  • Roughly speaking, PI patterns are sets (1) equipped with two strict partial orders—called precedence and inclusion—that interact with one another through laws called interactive transitivity and interactive irreflexivity, and (2) whose elements may be assigned attributes. PI patterns have a precisely-defined concept of generalization based on the existence of a pattern-preserving mapping from a general pattern to a more specific one, which provides a principled setting for exploring pattern generalization.
  • This well-founded approach to pattern generalization has deeper consequences. It leads to the fact that each nonempty set of PI patterns has a most specific generalization (MSG) that is again a PI pattern. Conceptually, an MSG plays a similar role to a least general generalization in inductive logic programming. Unfortunately, there is nothing canonical about an MSG of a set of patterns. However, it is a theorem that every nonempty finite set S of finite PI has a precisely defined best MSG, again a PI pattern, called a “minimal most specific generalization” (MNSG) of S. Moreover, an MMSG of a nonempty finite set of finite PI patterns is unique up to isomorphism.
  • The theory described above was developed independent of any particular approach to parsing in mind, and it is compatible with partial parsing approaches that do not attempt to fully analyze sentences. However, this theory can be used most advantageously with full parsers, such as English Slot Grammar (ESG), aided by statistical named-entity recognizers.
  • In the present invention, PI patterns are learned from small numbers of example sentences that contain Event mentions, and the learned PI patterns are then applied to other sentences in order to find new Event mentions. For this task, sentences in which named entity mentions are machine-labeled are parsed by ESG.
  • As shown by the exemplary flowchart of FIG. 8, the learning phase 800 proceeds as follows:
  • In step 801, the user picks out or creates a similar example sentence describing an Event of interest. It might be preferable, but is not necessary for the operation of the tool, that at least two similar sentences be presented.
  • In step 802, the user indicates the Entity mentions that fill the Event's argument roles in each example sentence.
  • In step 803, the parser will parse the sentence, and, in steps 804 and 805 the PI Pattern Learner 102 will compute PI patterns for each of the example sentences, including the information about the event of interest and compute the MMSG of the example PI patterns, which will necessarily contain what might be called a “generalized event”.
  • In step 806, the user will indicate either that another example sentence will be entered, by choosing the “Instance Definition” tab from the menu at the top of the tool display, thereby returning to step 801, or will exit the learning phase.
  • To find Event mentions in new, unseen sentences, the application phase will proceed in the following manner exemplarily shown in the process 900 of FIG. 9. The user enters this application phase by selecting the “Apply to Text” menu tab 705 (see FIG. 7).
  • In steps 901 and 902, a sentence is received from the document corpus and parsed into a parse tree. In steps 903, and 904, the PI Pattern Applier 106 computes the PI pattern of this sentence and, if the PI pattern is determined to be a special case of the MMSG of the learning phase by computing a pattern-preserving map from the more general pattern to the more specific one if there is one, determines the fillers for the argument roles of the event mentions in the sentence. If a pattern-preserving mapping was found in the previous step, in step 905 the PI Pattern Applier 106 determine the fillers for the argument roles of the Event mentions(s) in the sentence. In step 906, this sequence is continued for all sentences of the corpus.
  • FIG. 10 illustrates an exemplary block diagram 1000 of the major components that might comprise a computer tool for implementing the present invention. Graphical User Interface (GUI) module 1001 allows the user to provide inputs 1002 and provides the display information 1003 for the screen displays, such as discussed for FIGS. 5-7. Control module 1004 provides the control instructions to interconnect the various modules and to control the information passing therebetween. The functions of the linguistic analyzer 101, PI pattern learner 102, and PI pattern applier 106 have been previously discussed. Memory interface 1005 controls the flow of information between local memory, as well as the flow of information to a possible database such as might contain the corpus documents for either the learning examples 103 or the input documents 107 used in the search phase. Of course, it should be noted that, if the input documents for the search phase is directed to a search over an external network 1006, then the tool 1000 would include a network interface 1007, exemplarily shown in FIG. 10 as a submodule of the control module 1004.
  • An important note is made here that, although the present invention has been discussed as used to execute a search through a database or a series of databases or data sources, it should not be considered as being so limited. Thus, for example, the present invention might be utilized in environments in which there is no fixed database or data sources but, rather, a real-time data source. As one possible non-limiting example, the present invention might be used for searching real-time text data in an intelligence-gathering environment, wherein no discrete and readily-identifiable database exists.
  • Illustration of Technical Approach
  • For an actual demonstration of the approach of the present invention, a small corpus of real English sentences that contain many events of the same type, but still exhibit linguistic variability, was created. Since it was desirable not to be impeded by the problem of named entity coreference resolution, a corpus was used in which events of interest were fully described in single sentences.
  • This corpus, containing 280 sentences and headlines, was constructed from short news stories found on the ExecutiveSelect.com website. These stories dealt with executive position changes. Named entity mentions, specifically PERSON, LOCATION, OCCUPATION, and ORGANIZATION, were marked in the corpus by hand, and all sentences were parsed in the corpus with ESG.
  • In an experiment, two similar example sentences, containing instances of a relation that was called the ExecutiveChange relation. This is an “Event”, in ACE terminology. Automatic Content Extraction (ACE) is a Federal government NIST program having an objective to develop technology for extracting content from natural language text. ACE organizes annual competitive workshops to advance this objective.
  • In this experiment, the ExecutiveChange Event was defined to have five arguments called Person, Previous Position, Previous Company, New Position, and New Company. Thus, each ExecutiveChange Event could be viewed as a template with five slots to fill. Here is the first example sentence (see item 606 of FIG. 6):
      • William R. Graber, previously serving as vice president and chief financial officer of The Mead Corporation, has been named as senior vice president and chief financial officer of McKesson HBOC, Inc./(NYSE˜MCK) of San Francisco, Calif.
        The PI pattern derived from the parse tree of this sentence has 24 elements.
  • And here is the second example sentence (see item 607 of FIG. 6):
      • Brad Yopp, previously serving as director finance and administration of Ivex Packaging Corporation, has been named as treasurer of Research, Inc.
        The PI pattern derived from the parse tree of this sentence has 14 elements.
  • While these two example sentences resemble each other strongly (in particular, the verb forms are the same), it should be noted that there are significant differences. The parse tree of the first one, having 48 nodes, is almost twice as large as the parse tree of the second one, which has 28 nodes. The nodes that correspond to arguments for the ExecutiveChange Event were then marked in each sentence.
  • Note that this marking of nodes by a user requires no special knowledge of linguistics. In the first sentence, the slot fillers turned out to be, respectively, pattern elements 1, 5, 7, 12, and 16. This relation instance, including attributes of elements that come from the ESG parse tree, can be displayed as follows:
      • Known ExecutiveChange relation instance:
      • Person : 1 --> <hd = { “c= William R. Graber” “w= William R. Graber” }, ph = { “PERSON” “subj” “subj” “sg” “propn” “noun” } >
      • Previous Position : 5 --> <hd = { “w= vice president and chief financial officer” “c= vice president and chief financial officer” }, ph = { “OCCUPATION” “sg” “propn” “noun” “objprep” } >
      • Previous Company : 7 --> <hd ={ “c= Mead Corporation” “w= Mead Corporation” }, ph = { “ORGANIZATION” “sg” “propn” “noun” “objprep” } >
      • New Position : 12 --> <hd = { “w= senior vice president and chief financial officer” “c= senior vice president and chief financial officer” }, ph = { “OCCUPATION” “sg” “propn” “noun” “objprep” } >
      • New Company : 16 --> <hd = { “w= McKesson HBOC” “c= McKesson HBOC” }, ph = { “ORGANIZATION” “lconj” “sg” “propn” “noun” } >
  • The ExecutiveChange Event in the second sentence is similarly identified, and can be displayed as follows:
      • Known ExecutiveChange relation instance:
      • Person : 1 --> <hd = { “c= Brad Yopp” “w= Brad Yopp” }, ph= {“PERSON” “subj” “sg” “propn” “noun”} >
      • Previous Position : 5 --> <hd = { “w= director finance and administration” “c= director finance and administration” }, ph = { “OCCUPATION” “sg” “propn” “noun” “objprep”} >
      • Previous Company : 7 --> <hd = { “c= Ivex Packaging Corporation” “w= Ivex Packaging Corporation” }, ph = { “ORGANIZATION” “sg” “propn” “noun” “objprep” }>
      • New Position : 11 --> <hd = { “w= treasurer” “c= treasurer” }, ph = { “OCCUPATION” “sg” “propn” “noun” “objprep” } >
      • New Company : 13 --> <hd = { “c= Research, Inc” “w= Research, Inc” }, ph = { “ORGANIZATION” “sg” “propn” “noun” “objprep” } >
        The product pattern of these 2 PI patterns has 336 elements.
  • Next, a Java® package which can be applied to any PI pattern generalization problem, computed the MMSG of these two patterns, which is not a parse tree, and in which there is a generalized ExecutiveChange relation instance that says nothing specific about the text that appears in the slot fillers. The MMSG of these 2 patterns has 15 elements.
  • The constructed generalized ExecutiveChange relation instance is partially described as:
      • Person : 1 --> <hd = { }, ph = { “PERSON” “subj” “sg” “propn” “noun”} >
      • Previous Position : 5 --> <hd = { }, ph = { “OCCUPATION” “sg” “propn” “objprep” “noun”} >
      • Previous Company : 7 --> <hd = { }, ph = { “ORGANIZATION” “sg” “propn” “objprep” “noun” } >
      • New Position : 11 --> <hd = { }, ph = { “OCCUPATION” “sg” “propn” “objprep” “noun” } >
      • New Company : 13 --> <hd = { }, ph = { “ORGANIZATION” “sg” “propn” “noun” } >
  • Note that the MMSG has more elements than the smaller of the two example patterns that generated it. This is an indication of the fact that one does not get an MMSG by deleting pattern elements in one PI pattern that somehow do not match any element in another PI pattern. When a program applies this pattern to the corpus, thirteen new mentions of the ExecutiveChange Event will be discovered in a few seconds, such as:
      • Lance J. Bennett, previously serving as vice president and associate general counsel of The Dime Savings Bank of New York, has been named as vice president and general counsel of Dime Community Bancshares, Inc.
  • The new ExecutiveChange relation instance that was found is:
      • Person : 1 --> hd = { “w= Lance J. Bennett” “c=Lance J. Bennett” }
      • Previous Position : 5 > hd = { “c=vice president and associate general counsel” “w= vice president and associate general 5 counsel” }
      • Previous Company : 7 --> hd = { “w= Dime Savings Bank of New York” “c= Dime Savings Bank of New York” }
      • New Position : 12 --> hd = { “c= vice president and general counsel” w= vice president and general counsel” }
      • New Company: 14 --> hd = { “c=Dime Community Bancshares, Inc” “w= Dime Community Bancshares, Inc” }
  • Although the ExecutiveChange Event slot fillers that come from sentences containing the verb form “has been named” could have been found by writing a simple ad hoc program, the important points to keep in kind are that this processing can be done in a completely general way, and no knowledge of linguistics is needed on the part of a user to define PI patterns that can then be used to extract knowledge from text. It is noted that, in the very limited tests carried out to test the present invention, no errors were made in finding arguments for the ExecutiveChange Event.
  • It is also noted that the above demonstration shows that it is possible for one of the example sentences to have an active verb and the other to have a passive verb. Additionally, the technique can handle the absence of some relation arguments from an example used to generate a PI pattern.
  • Brief Description of the Mathematical Theory of Precedence Inclusion Patterns
  • The following very brief discussion of the mathematical theory underlying the present invention is further discussed in a paper entitled “Patterns Based on Multiple Interacting Partial Orders” by one of the co-inventors, Frank J. Oles of IBM T.J. Watson Research Center, Yorktown Heights, N.Y. 10598, as presented in the above-referenced Provisional Application 60/586,877.
  • It is first noted that the mathematical theory encompasses more than the method discussed above to extract information from unstructured text, since the theory additionally applies to extraction of information from images. Thus, the present invention can be viewed as a concrete implementation of a somewhat narrow, specific application of this new mathematical theory as it relates to text. That is, the tool of the present invention is not currently implemented to apply the precedence inclusion pattern theory to image analysis, and it is expected that several problems need to be solved before the theory can be embodied in a tool for images. However, the interested reader is invited to read this paper for more details of this mathematical theory than is discussed below.
  • The referenced paper outlines a theory of patterns where the patterns are based on multiple interacting strict partial orders. For instance, a pattern extracted from a video may have as its elements some picture elements occurring in individual frames of the video. There are four natural strict partial orders relating those elements: elements may be ordered by the ordering of the frames in which they occur, and elements within a single frame may be above one another, to the left of one another, and included within one another.
  • Other examples come from parsing text (e.g., the subject of the present invention), where linguistic entities in a parse tree may precede one another or they may contain one another. In fact, categories of what is referred to in the paper as “2-patterns” include among their objects constituent structure trees, as they are normally defined in computational linguistics.
  • As mentioned previously, the theory in this paper was motivated by current problems of relational learning, an important kind of inductive learning in which one wishes, from known training instances of related elements of structures, to create general rules for identifying elements of other structures that bear the same relation to one another.
  • For instance, one may wish to learn from text examples patterns expressing the fact that a disease has a symptom (a binary relation) or that a person has a position in a company (a ternary relation). As another example, one may want to learn the properties that some nucleotide sequences have in common as well as learning out how to pick out a particular subsequence of interest (a unary relation).
  • The general supervised learning problem of classification can be cast as learning a 0-ary relation. This inductive learning approach of this paper is called category-theoretic inductive learning since the notion of generalization employed is based on morphisms between structured objects. An ordered pair [<, ⊃] of binary relations on a set is said to be interactively transitive if both < and ⊃ are transitive and, for all x, y, zεP,
      • 1. x<y and y⊃z implies x<x, and
      • 2. y⊃x and y<z implies x<z.
  • Interactive transitivity is an extension of the concept of a transitive, binary relation to an ordered pair of binary relations. By reading as the symbology “<” as meaning “precedes” and the symbology “⊃” as meaning “includes,” the intuitive content of these axioms may become more clear to the reader.
  • For instance, in the domain of two-dimensional images, the ordered pair of relations [is to the left of, contains] is interactively transitive, as also would be the ordered pair [is above, contains].
  • Roughly speaking, a precedence-inclusion pattern is a set equipped with a strictly partially ordered set of strict partial orders, along with some additional structure, in which the strict partial order on the strict partial orders is taken to assert that each related pair of strict partial orders obeys the axioms of interactive transitivity. A precise definition starts in an algebraic style. A pattern signature an ordered triple Σ=(O, A, L) in which:
      • 1. O, the order symbol set of Σ, is a strictly partially ordered set of binary relation symbols, each of which is intended to be interpreted as a strict partial order on a set;
      • 2. A, the argument name set of Σ, is a set whose elements name the arguments for some A-ary relation of interest, instances of which may be found in patterns; and
      • 3. L, the property poset of Σ, is a bounded complete poset of labels that may be attached to elements of structures.
        Definition
  • Let τ=(O, A, L) be a pattern signature. It is said that a set is a Σ-pattern when every σεO has an interpretation
    Figure US20060009966A1-20060112-P00900
    σ,P as a strict partial order on P, along with a partial function αP: P→L, called the argument namingfunction, and a total function ΛP: P→L, called the labeling function, such that σ<τ implies that the ordered pair of relations [
    Figure US20060009966A1-20060112-P00900
    σ,P,
    Figure US20060009966A1-20060112-P00900
    τ,P] is interactively transitive.
  • When Σ is clear from context, a Σ-pattern is called a precedence-inclusion pattern. Thus, when the order symbol set O is empty, Σ-patterns are just sets with some additional structure. When the order symbol set is one-element set, then Σ-patterns are strictly partially ordered sets with some additional structure.
  • More interesting examples arise when the order symbol set is nontrivial. Examples of precedence-inclusion patterns in which the order symbol set has arbitrary finite depth can be constructed.
  • A Σ-pattern Q is a generalization of a Σ-pattern if there is a pattern-preserving map, i.e., a morphism in the category of Σ-patterns, from Q to P.
  • The reader can now guess at the definition of a most specific generalization (msg) of a set of patterns, which corresponds to a least general generalization (lgg) in inductive logic programming. Like lgg's, msg's are not unique, although products of patterns give (typically very large) examples of them. The problem with a large msg is that it would be computationally hard to test if another pattern is a specialization of it.
  • A minimal most specific generalization of a set P of patterns is an msg of P no subpattern of which is an msg of P. These are the kinds of generalizations that are desired. A retraction of a precedence-inclusion pattern P is an idempotent endomorphism r: P→P, and the set of fixed points of a retraction defines a pattern called a retract of P. A pattern having no proper retracts is said to be fully retracted.
  • Here is the main theorem. For the finite case, it covers the existence and uniqueness of the minimal most specific generalization, and, implicitly, tells how to compute it.
  • Theorem
  • Let I be a nonempty finite index set and let P={Pi|iεI} be an I-indexed set of finite Σ-patterns.
      • 1. There exists a minimal most specific generalization M of P.
      • 2. M is finite and fully retracted.
      • 3. Any minimal most specific generalization of P is isomorphic to M.
      • 4. Any finite most specific generalization Q of P has a retraction r: Q→Q whose image is isomorphic to M.
  • Below is an exemplary explicit description of a simple procedure that is guaranteed to return the minimal most specific generalization of a nonempty finite set {P1, P2, . . . , Pn} of finite Σ-patterns.
  • Minimal Most Specific Generalization Procedure
    M:=P 1 ×P 2 × . . . ×P n;
      • while there exists a proper retract Q of M
        do M:=Q;
      • return M;
        Hardware Implementation
  • FIG. 11 illustrates a typical hardware configuration of an information handling/computer system in accordance with the invention and which preferably has at least one processor or central processing unit (CPU) 1111.
  • The CPUs 1111 are interconnected via a system bus 1112 to a random access memory (RAM) 1114, read-only memory (ROM) 1116, input/output (I/O) adapter 1118 (for connecting peripheral devices such as disk units 1121 and tape drives 1140 to the bus 1112), user interface adapter 1122 (for connecting a keyboard 1124, mouse 1126, speaker 1128, microphone 1132, and/or other user interface device to the bus 1112), a communication adapter 1134 for connecting an information handling system to a data processing network, the Internet, an Intranet, a personal area network (PAN), etc., and a display adapter 1136 for connecting the bus 1112 to a display device 1138 and/or printer 1139 (e.g., a digital printer or the like).
  • In addition to the hardware/software environment described above, a different aspect of the invention includes a computer-implemented method for performing the above method. As an example, this method may be implemented in the particular environment discussed above.
  • Such a method may be implemented, for example, by operating a computer, as embodied by a digital data processing apparatus, to execute a sequence of machine-readable instructions. These instructions may reside in various types of signal-bearing media.
  • Thus, this aspect of the present invention is directed to a programmed product, comprising signal-bearing media tangibly embodying a program of machine-readable instructions executable by a digital data processor incorporating the CPU 1111 and hardware above, to perform the method of the invention.
  • This signal-bearing media may include, for example, a RAM contained within the CPU 1111, as represented by the fast-access storage for example. Alternatively, the instructions may be contained in another signal-bearing media, such as a magnetic data storage diskette 1200 (FIG. 12), directly or indirectly accessible by the CPU 1111.
  • Whether contained in the diskette 1200, the computer/CPU 1111, or elsewhere, the instructions may be stored on a variety of machine-readable data storage media, such as DASD storage (e.g., a conventional “hard drive” or a RAID array), magnetic tape, electronic read-only memory (e.g., ROM, EPROM, or EEPROM), an optical storage device (e.g. CD-ROM, WORM, DVD, digital optical tape, etc.), paper “punch” cards, or other suitable signal-bearing media including transmission media such as digital and analog and communication links and wireless. In an illustrative embodiment of the invention, the machine-readable instructions may comprise software object code.
  • Software Implementation
  • Although, in terms of software implementation, the present invention has been demonstrated in Java® for pairs of finite 2-pattern, the resulting minimal msg's can readily be used for discovering new instances of relations. Moreover, it is not intended that the present invention be limited to 2-patterns or to implementation in Java®.
  • It is noted that, although the prototype has been demonstrated for text in the English language, it should be apparent that one of ordinary skill in the art, after taking the disclosure as a whole, would be able to adapt the present invention to text in other languages. Thus, there is no intent that the method described above be limited to any specific language.
  • Using the Present Invention as a Service
  • In yet another aspect of the present invention, it is noted that the present invention can be used as a basis for a service or business method. In this aspect, the present invention might be, for example, implemented as a search method for specific databases or, more generally, for any number of data sources available through a computer network, such as the Internet.
  • In this aspect, the present invention is intended as covering, not only the tool itself that executes the methods described above, but also a service made available for using this method by others. Thus, for example, an entity might provide the service of conducting searches of databases or the Internet, using the method of the present invention. The present invention is intended as including this exploitation of executing the above-described methods by providing a service to others to conduct searches using these methods. Another service related to the present invention might be based on making a tool available to others so that these clients or customers can themselves use the tool to conduct their own search.
  • Specific Examples of Searches Possible with the Present Invention
  • It should go without statement that the present invention can be used for almost any type of search, so that the few example scenarios discussed above should not be considered limiting.
  • However, along this line, while there is no reason to limit the present invention to specific examples wherein it might be used, a number of non-limiting examples for which it might be employed might include: a web search, wherein a user picks examples and searches for the same relations; intelligence gathering by intelligence agencies; searches by financial analysts, looking for facts, press releases, or SEC filings; or extracting information from clinical notes of medical practitioners.
  • The searches might be done on propriety or publicly available databases, a computer network, including the Internet, or might even be done in real-time.
  • While the invention has been described in terms of an exemplary embodiment, those skilled in the art will recognize that the invention can be practiced with modification within the spirit and scope of the appended claims.
  • Further, it is noted that Applicants' intent is to encompass equivalents of all claim elements, even if amended later during prosecution.

Claims (22)

1. A method of extracting information from text, said method comprising:
parsing an input sample of text to form a parse tree; and
receiving user inputs to define a machine-labeled learning pattern from said parse tree.
2. The method of claim 1, further comprising:
calculating a generalization of said learning pattern that is not also a parse tree.
3. The method of claim 1, wherein machine-labeled learning pattern comprises a precedence inclusion pattern wherein elements in said learning pattern are defined in a precedence relation and in an inclusion relation.
4. The method of claim 3, wherein said input sample comprises a first input sample, said parse tree comprises a first parse tree, and said learning pattern comprises a first learning pattern, said method further comprising:
parsing at least one more input sample of text to form therefrom a parse tree;
for each said at least one more input sample parse tree, defining therefrom a learning pattern; and
calculating a generalization of said learning patterns.
5. The method of claim 4, wherein each said learning pattern comprises a precedence inclusion pattern wherein elements in said learning pattern are defined in a precedence relation and in an inclusion relation.
6. The method of claim 5, wherein said generalization also comprises a precedence inclusion pattern.
7. The method of claim 6, wherein said precedence inclusion pattern of said generalization comprises a most specific generalization (MSG).
8. The method of claim 7, further comprising:
calculating a Minimal Most Specific Generalization (MMSG) of all of said learning samples.
9. The method of claim 2, further comprising:
comparing said learning pattern with an unknown text.
10. The method of claim 8, further comprising:
comparing said learning patterns with an unknown text
11. The method of claim 10, wherein said comparing comprises:
parsing each said unknown text to form a parse tree;
calculating a generalization of said parse tree, said generalization forming a precedence inclusion pattern; and
using said MMSG to calculate a similarity of said unknown text to said learning patterns.
12. The method of claim 9, wherein said comparing comprises:
parsing each said unknown text to form a parse tree;
calculating a generalization of said parse tree, said generalization forming a precedence inclusion pattern; and
calculating a similarity of said generalization of said parse tree of said unknown text with said generalization of said learning pattern.
13. An apparatus for relational learning, said apparatus comprising:
a generator for developing a precedence inclusion (PI) pattern of a learning sample, wherein elements in said learning sample are machine-labeled to define a precedence relation and an inclusion relation.
14. The apparatus of claim 13, further comprising:
a graphical user interface (GUI) to permit a user to provide inputs used for said developing said PI pattern.
15. The apparatus of claim 13, further comprising:
a comparison module for applying said PI pattern to unseen text and determining a similarity therebetween.
16. The apparatus of claim 15, wherein said generator further calculates a Minimal Most Specific Generalization (MMSG) of all learning samples entered and said comparison is based on said MMSG.
17. A signal-bearing medium tangibly embodying a program of machine-readable instructions executable by a digital processing apparatus to perform a method of relational learning, said machine-readable instructions comprising:
a precedence inclusion (PI) pattern learning module for generating a PI pattern of a learning sample wherein elements in said learning sample are machine-labeled to define a precedence relation and an inclusion relation.
18. The signal-bearing medium of claim 17, further comprising:
a graphical user interface (GUI) to permit a user to provide inputs to define said PI pattern for each said learning sample.
19. The signal-bearing medium of claim 17, wherein said PI pattern learning module further calculates a Minimal Most Specific Generalization (MMSG) of all learning samples entered, said machine-readable instructions further comprising:
a PI application module for comparing unseen text with said learning samples, said comparing based on said MMSG.
20. A method of searching unseen text, said method comprising at least one of:
conducting a search of unseen text by developing a precedence inclusion (PI) pattern of at least one learning sample and using said PI pattern for comparison with unseen text; and
providing a computerized tool to a user for said conducting said search.
21. An apparatus for extracting information from text, said apparatus comprising:
means for parsing an input sample of text to form a parsed tree; and
means for receiving user inputs to define a machine-labeled learning pattern from said parsed tree.
22. A computerized tool for extracting information from text, said computerized tool comprising:
a precedence inclusion (PI) pattern learning module for generating a PI pattern of a learning sample wherein elements in said learning sample are machine-labeled to define a precedence relation and an inclusion relation.
US10/979,162 2004-07-12 2004-11-03 Method and system for extracting information from unstructured text using symbolic machine learning Abandoned US20060009966A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US10/979,162 US20060009966A1 (en) 2004-07-12 2004-11-03 Method and system for extracting information from unstructured text using symbolic machine learning
US12/507,866 US8140323B2 (en) 2004-07-12 2009-07-23 Method and system for extracting information from unstructured text using symbolic machine learning

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US58687704P 2004-07-12 2004-07-12
US10/979,162 US20060009966A1 (en) 2004-07-12 2004-11-03 Method and system for extracting information from unstructured text using symbolic machine learning

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US12/507,866 Continuation US8140323B2 (en) 2004-07-12 2009-07-23 Method and system for extracting information from unstructured text using symbolic machine learning

Publications (1)

Publication Number Publication Date
US20060009966A1 true US20060009966A1 (en) 2006-01-12

Family

ID=35542457

Family Applications (2)

Application Number Title Priority Date Filing Date
US10/979,162 Abandoned US20060009966A1 (en) 2004-07-12 2004-11-03 Method and system for extracting information from unstructured text using symbolic machine learning
US12/507,866 Active 2025-05-09 US8140323B2 (en) 2004-07-12 2009-07-23 Method and system for extracting information from unstructured text using symbolic machine learning

Family Applications After (1)

Application Number Title Priority Date Filing Date
US12/507,866 Active 2025-05-09 US8140323B2 (en) 2004-07-12 2009-07-23 Method and system for extracting information from unstructured text using symbolic machine learning

Country Status (1)

Country Link
US (2) US20060009966A1 (en)

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070250505A1 (en) * 2006-04-25 2007-10-25 Sbc Knowledge Ventures, L.P. Method and apparatus for defining a workflow
US20070250613A1 (en) * 2006-04-25 2007-10-25 Sbc Knowledge Ventures, L.P. Method and apparatus for configuring a workflow
US20070250822A1 (en) * 2006-04-25 2007-10-25 Sbc Knowledge Ventures, L.P. Method and apparatus for importing content in a user-defined workflow
US20080133443A1 (en) * 2006-11-30 2008-06-05 Bohannon Philip L Methods and Apparatus for User-Guided Inference of Regular Expressions for Information Extraction
US20090198646A1 (en) * 2008-01-31 2009-08-06 International Business Machines Corporation Systems, methods and computer program products for an algebraic approach to rule-based information extraction
US20100082331A1 (en) * 2008-09-30 2010-04-01 Xerox Corporation Semantically-driven extraction of relations between named entities
US20100121631A1 (en) * 2008-11-10 2010-05-13 Olivier Bonnet Data detection
US20100293451A1 (en) * 2006-06-21 2010-11-18 Carus Alwin B An apparatus, system and method for developing tools to process natural language text
US20110035210A1 (en) * 2009-08-10 2011-02-10 Benjamin Rosenfeld Conditional random fields (crf)-based relation extraction system
US20110131216A1 (en) * 2006-09-08 2011-06-02 International Business Machines Corporation Automatically linking documents with relevant structured information
US20110144971A1 (en) * 2009-12-16 2011-06-16 Computer Associates Think, Inc. System and method for sentiment analysis
WO2011134141A1 (en) * 2010-04-27 2011-11-03 Hewlett-Packard Development Company,L.P. Method of extracting named entity
US8738360B2 (en) 2008-06-06 2014-05-27 Apple Inc. Data detection of a character sequence having multiple possible data types
US9043197B1 (en) * 2006-07-14 2015-05-26 Google Inc. Extracting information from unstructured text using generalized extraction patterns
US20150324436A1 (en) * 2012-12-28 2015-11-12 Hitachi, Ltd. Data processing system and data processing method
WO2016059505A1 (en) * 2014-10-14 2016-04-21 Uab "Locatory.Com" A system and a method for recognition of aerospace parts in unstructured text
US10558760B2 (en) 2017-07-28 2020-02-11 International Business Machines Corporation Unsupervised template extraction
US10838953B1 (en) 2008-07-21 2020-11-17 NetBase Solutions, Inc. Method and apparatus for frame based search
US10872082B1 (en) 2011-10-24 2020-12-22 NetBase Solutions, Inc. Methods and apparatuses for clustered storage of information
US10885324B2 (en) * 2019-04-11 2021-01-05 Adp, Llc Agency notice processing system
US11055295B1 (en) * 2010-04-22 2021-07-06 NetBase Solutions, Inc. Method and apparatus for determining search result demographics
US11210473B1 (en) 2020-03-12 2021-12-28 Yseop Sa Domain knowledge learning techniques for natural language generation
JP2022122029A (en) * 2021-02-09 2022-08-22 株式会社東芝 Data processing device, data processing method, and data processing program
US11449687B2 (en) 2019-05-10 2022-09-20 Yseop Sa Natural language text generation using semantic objects
US11494560B1 (en) * 2020-01-30 2022-11-08 Act, Inc. System and methodology for computer-facilitated development of reading comprehension test items through passage mapping
US11501088B1 (en) 2020-03-11 2022-11-15 Yseop Sa Techniques for generating natural language text customized to linguistic preferences of a user
US11599580B2 (en) * 2018-11-29 2023-03-07 Tata Consultancy Services Limited Method and system to extract domain concepts to create domain dictionaries and ontologies

Families Citing this family (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8977953B1 (en) * 2006-01-27 2015-03-10 Linguastat, Inc. Customizing information by combining pair of annotations from at least two different documents
US8731954B2 (en) 2006-03-27 2014-05-20 A-Life Medical, Llc Auditing the coding and abstracting of documents
US8762834B2 (en) * 2006-09-29 2014-06-24 Altova, Gmbh User interface for defining a text file transformation
US8671341B1 (en) 2007-01-05 2014-03-11 Linguastat, Inc. Systems and methods for identifying claims associated with electronic text
US8682823B2 (en) 2007-04-13 2014-03-25 A-Life Medical, Llc Multi-magnitudinal vectors with resolution based on source vector features
US7908552B2 (en) 2007-04-13 2011-03-15 A-Life Medical Inc. Mere-parsing with boundary and semantic driven scoping
US9946846B2 (en) * 2007-08-03 2018-04-17 A-Life Medical, Llc Visualizing the documentation and coding of surgical procedures
WO2011137935A1 (en) 2010-05-07 2011-11-10 Ulysses Systems (Uk) Limited System and method for identifying relevant information for an enterprise
US8903128B2 (en) * 2011-02-16 2014-12-02 Siemens Aktiengesellschaft Object recognition for security screening and long range video surveillance
US8930380B1 (en) * 2011-06-30 2015-01-06 Sumo Logic Automatic parser generation
US9514257B1 (en) 2011-10-30 2016-12-06 Lockheed Martin Corporation Event visualization based on unstructured data
US20130332450A1 (en) * 2012-06-11 2013-12-12 International Business Machines Corporation System and Method for Automatically Detecting and Interactively Displaying Information About Entities, Activities, and Events from Multiple-Modality Natural Language Sources
US20140082003A1 (en) * 2012-09-17 2014-03-20 Digital Trowel (Israel) Ltd. Document mining with relation extraction
US9471559B2 (en) * 2012-12-10 2016-10-18 International Business Machines Corporation Deep analysis of natural language questions for question answering system
US9348815B1 (en) 2013-06-28 2016-05-24 Digital Reasoning Systems, Inc. Systems and methods for construction, maintenance, and improvement of knowledge representations
US10191893B2 (en) 2013-07-22 2019-01-29 Open Text Holdings, Inc. Information extraction and annotation systems and methods for documents
US8856642B1 (en) 2013-07-22 2014-10-07 Recommind, Inc. Information extraction and annotation systems and methods for documents
US10541053B2 (en) 2013-09-05 2020-01-21 Optum360, LLCq Automated clinical indicator recognition with natural language processing
US10133727B2 (en) 2013-10-01 2018-11-20 A-Life Medical, Llc Ontologically driven procedure coding
US9501525B2 (en) * 2014-11-05 2016-11-22 International Business Machines Corporation Answer sequence evaluation
US10324965B2 (en) 2014-12-30 2019-06-18 International Business Machines Corporation Techniques for suggesting patterns in unstructured documents
US10169423B2 (en) 2016-01-06 2019-01-01 International Business Machines Corporation Ranking answers in ground truth of a question-answering system
US10073834B2 (en) 2016-02-09 2018-09-11 International Business Machines Corporation Systems and methods for language feature generation over multi-layered word representation
US10528661B2 (en) 2016-02-11 2020-01-07 International Business Machines Corporation Evaluating parse trees in linguistic analysis
US10726054B2 (en) 2016-02-23 2020-07-28 Carrier Corporation Extraction of policies from natural language documents for physical access control
US11048762B2 (en) 2018-03-16 2021-06-29 Open Text Holdings, Inc. User-defined automated document feature modeling, extraction and optimization
US10762142B2 (en) 2018-03-16 2020-09-01 Open Text Holdings, Inc. User-defined automated document feature extraction and optimization
US10762301B1 (en) * 2018-09-04 2020-09-01 Michael Dudley Johnson Methods and systems for generating linguistic rules
US11610277B2 (en) 2019-01-25 2023-03-21 Open Text Holdings, Inc. Seamless electronic discovery system with an enterprise data portal
US11645513B2 (en) * 2019-07-03 2023-05-09 International Business Machines Corporation Unary relation extraction using distant supervision
US20220207384A1 (en) * 2020-12-30 2022-06-30 International Business Machines Corporation Extracting Facts from Unstructured Text

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5519608A (en) * 1993-06-24 1996-05-21 Xerox Corporation Method for extracting from a text corpus answers to questions stated in natural language by using linguistic analysis and hypothesis generation
US5864788A (en) * 1992-09-25 1999-01-26 Sharp Kabushiki Kaisha Translation machine having a function of deriving two or more syntaxes from one original sentence and giving precedence to a selected one of the syntaxes
US5963742A (en) * 1997-09-08 1999-10-05 Lucent Technologies, Inc. Using speculative parsing to process complex input data
US6246977B1 (en) * 1997-03-07 2001-06-12 Microsoft Corporation Information retrieval utilizing semantic representation of text and based on constrained expansion of query words
US20020169596A1 (en) * 2001-05-04 2002-11-14 Brill Eric D. Method and apparatus for unsupervised training of natural language processing units
US6556983B1 (en) * 2000-01-12 2003-04-29 Microsoft Corporation Methods and apparatus for finding semantic information, such as usage logs, similar to a query using a pattern lattice data space
US20030144978A1 (en) * 2002-01-17 2003-07-31 Zeine Hatem I. Automated learning parsing system
US20030212544A1 (en) * 2002-05-10 2003-11-13 Alejandro Acero System for automatically annotating training data for a natural language understanding system
US6675159B1 (en) * 2000-07-27 2004-01-06 Science Applic Int Corp Concept-based search and retrieval system
US20040044519A1 (en) * 2002-08-30 2004-03-04 Livia Polanyi System and method for summarization combining natural language generation with structural analysis
US20040111253A1 (en) * 2002-12-10 2004-06-10 International Business Machines Corporation System and method for rapid development of natural language understanding using active learning
US20040176945A1 (en) * 2003-03-06 2004-09-09 Nagoya Industrial Science Research Institute Apparatus and method for generating finite state transducer for use in incremental parsing
US20040220797A1 (en) * 2003-05-01 2004-11-04 Microsoft Corporation Rules-based grammar for slots and statistical model for preterminals in natural language understanding system
US20040243394A1 (en) * 2003-05-28 2004-12-02 Oki Electric Industry Co., Ltd. Natural language processing apparatus, natural language processing method, and natural language processing program
US7003445B2 (en) * 2001-07-20 2006-02-21 Microsoft Corporation Statistically driven sentence realizing method and apparatus
US7191119B2 (en) * 2002-05-07 2007-03-13 International Business Machines Corporation Integrated development tool for building a natural language understanding application
US7249117B2 (en) * 2002-05-22 2007-07-24 Estes Timothy W Knowledge discovery agent system and method

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10214203A (en) * 1997-01-29 1998-08-11 Nec Corp Information processor
US6901402B1 (en) * 1999-06-18 2005-05-31 Microsoft Corporation System for improving the performance of information retrieval-type tasks by identifying the relations of constituents
US20040205482A1 (en) * 2002-01-24 2004-10-14 International Business Machines Corporation Method and apparatus for active annotation of multimedia content
US7219054B1 (en) * 2003-04-04 2007-05-15 At&T Corp. Systems and methods for generating an annotation guide
US7440890B2 (en) * 2003-12-19 2008-10-21 Xerox Corporation Systems and methods for normalization of linguisitic structures

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5864788A (en) * 1992-09-25 1999-01-26 Sharp Kabushiki Kaisha Translation machine having a function of deriving two or more syntaxes from one original sentence and giving precedence to a selected one of the syntaxes
US5519608A (en) * 1993-06-24 1996-05-21 Xerox Corporation Method for extracting from a text corpus answers to questions stated in natural language by using linguistic analysis and hypothesis generation
US6246977B1 (en) * 1997-03-07 2001-06-12 Microsoft Corporation Information retrieval utilizing semantic representation of text and based on constrained expansion of query words
US5963742A (en) * 1997-09-08 1999-10-05 Lucent Technologies, Inc. Using speculative parsing to process complex input data
US6556983B1 (en) * 2000-01-12 2003-04-29 Microsoft Corporation Methods and apparatus for finding semantic information, such as usage logs, similar to a query using a pattern lattice data space
US6675159B1 (en) * 2000-07-27 2004-01-06 Science Applic Int Corp Concept-based search and retrieval system
US20020169596A1 (en) * 2001-05-04 2002-11-14 Brill Eric D. Method and apparatus for unsupervised training of natural language processing units
US7003445B2 (en) * 2001-07-20 2006-02-21 Microsoft Corporation Statistically driven sentence realizing method and apparatus
US20030144978A1 (en) * 2002-01-17 2003-07-31 Zeine Hatem I. Automated learning parsing system
US7191119B2 (en) * 2002-05-07 2007-03-13 International Business Machines Corporation Integrated development tool for building a natural language understanding application
US20030212544A1 (en) * 2002-05-10 2003-11-13 Alejandro Acero System for automatically annotating training data for a natural language understanding system
US7249117B2 (en) * 2002-05-22 2007-07-24 Estes Timothy W Knowledge discovery agent system and method
US20040044519A1 (en) * 2002-08-30 2004-03-04 Livia Polanyi System and method for summarization combining natural language generation with structural analysis
US20040111253A1 (en) * 2002-12-10 2004-06-10 International Business Machines Corporation System and method for rapid development of natural language understanding using active learning
US20040176945A1 (en) * 2003-03-06 2004-09-09 Nagoya Industrial Science Research Institute Apparatus and method for generating finite state transducer for use in incremental parsing
US20040220797A1 (en) * 2003-05-01 2004-11-04 Microsoft Corporation Rules-based grammar for slots and statistical model for preterminals in natural language understanding system
US20040243394A1 (en) * 2003-05-28 2004-12-02 Oki Electric Industry Co., Ltd. Natural language processing apparatus, natural language processing method, and natural language processing program

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070250613A1 (en) * 2006-04-25 2007-10-25 Sbc Knowledge Ventures, L.P. Method and apparatus for configuring a workflow
US20070250822A1 (en) * 2006-04-25 2007-10-25 Sbc Knowledge Ventures, L.P. Method and apparatus for importing content in a user-defined workflow
US20070250505A1 (en) * 2006-04-25 2007-10-25 Sbc Knowledge Ventures, L.P. Method and apparatus for defining a workflow
US8131756B2 (en) * 2006-06-21 2012-03-06 Carus Alwin B Apparatus, system and method for developing tools to process natural language text
US20100293451A1 (en) * 2006-06-21 2010-11-18 Carus Alwin B An apparatus, system and method for developing tools to process natural language text
US9043197B1 (en) * 2006-07-14 2015-05-26 Google Inc. Extracting information from unstructured text using generalized extraction patterns
US20110131216A1 (en) * 2006-09-08 2011-06-02 International Business Machines Corporation Automatically linking documents with relevant structured information
US8126892B2 (en) * 2006-09-08 2012-02-28 International Business Machines Corporation Automatically linking documents with relevant structured information
US20080133443A1 (en) * 2006-11-30 2008-06-05 Bohannon Philip L Methods and Apparatus for User-Guided Inference of Regular Expressions for Information Extraction
US20090198646A1 (en) * 2008-01-31 2009-08-06 International Business Machines Corporation Systems, methods and computer program products for an algebraic approach to rule-based information extraction
US9454522B2 (en) 2008-06-06 2016-09-27 Apple Inc. Detection of data in a sequence of characters
US8738360B2 (en) 2008-06-06 2014-05-27 Apple Inc. Data detection of a character sequence having multiple possible data types
US11886481B2 (en) 2008-07-21 2024-01-30 NetBase Solutions, Inc. Method and apparatus for frame-based search and analysis
US10838953B1 (en) 2008-07-21 2020-11-17 NetBase Solutions, Inc. Method and apparatus for frame based search
US20100082331A1 (en) * 2008-09-30 2010-04-01 Xerox Corporation Semantically-driven extraction of relations between named entities
US8370128B2 (en) * 2008-09-30 2013-02-05 Xerox Corporation Semantically-driven extraction of relations between named entities
US8489388B2 (en) * 2008-11-10 2013-07-16 Apple Inc. Data detection
US9489371B2 (en) 2008-11-10 2016-11-08 Apple Inc. Detection of data in a sequence of characters
US20100121631A1 (en) * 2008-11-10 2010-05-13 Olivier Bonnet Data detection
US20110035210A1 (en) * 2009-08-10 2011-02-10 Benjamin Rosenfeld Conditional random fields (crf)-based relation extraction system
US8843362B2 (en) * 2009-12-16 2014-09-23 Ca, Inc. System and method for sentiment analysis
US20110144971A1 (en) * 2009-12-16 2011-06-16 Computer Associates Think, Inc. System and method for sentiment analysis
US11055295B1 (en) * 2010-04-22 2021-07-06 NetBase Solutions, Inc. Method and apparatus for determining search result demographics
US20130204835A1 (en) * 2010-04-27 2013-08-08 Hewlett-Packard Development Company, Lp Method of extracting named entity
WO2011134141A1 (en) * 2010-04-27 2011-11-03 Hewlett-Packard Development Company,L.P. Method of extracting named entity
US10872082B1 (en) 2011-10-24 2020-12-22 NetBase Solutions, Inc. Methods and apparatuses for clustered storage of information
US11681700B1 (en) 2011-10-24 2023-06-20 NetBase Solutions, Inc. Methods and apparatuses for clustered storage of information
US20150324436A1 (en) * 2012-12-28 2015-11-12 Hitachi, Ltd. Data processing system and data processing method
WO2016059505A1 (en) * 2014-10-14 2016-04-21 Uab "Locatory.Com" A system and a method for recognition of aerospace parts in unstructured text
US10572601B2 (en) 2017-07-28 2020-02-25 International Business Machines Corporation Unsupervised template extraction
US10558760B2 (en) 2017-07-28 2020-02-11 International Business Machines Corporation Unsupervised template extraction
US11599580B2 (en) * 2018-11-29 2023-03-07 Tata Consultancy Services Limited Method and system to extract domain concepts to create domain dictionaries and ontologies
US10885324B2 (en) * 2019-04-11 2021-01-05 Adp, Llc Agency notice processing system
US11449687B2 (en) 2019-05-10 2022-09-20 Yseop Sa Natural language text generation using semantic objects
US11809832B2 (en) 2019-05-10 2023-11-07 Yseop Sa Natural language text generation using semantic objects
US11494560B1 (en) * 2020-01-30 2022-11-08 Act, Inc. System and methodology for computer-facilitated development of reading comprehension test items through passage mapping
US11501088B1 (en) 2020-03-11 2022-11-15 Yseop Sa Techniques for generating natural language text customized to linguistic preferences of a user
US11210473B1 (en) 2020-03-12 2021-12-28 Yseop Sa Domain knowledge learning techniques for natural language generation
JP2022122029A (en) * 2021-02-09 2022-08-22 株式会社東芝 Data processing device, data processing method, and data processing program

Also Published As

Publication number Publication date
US8140323B2 (en) 2012-03-20
US20090287476A1 (en) 2009-11-19

Similar Documents

Publication Publication Date Title
US8140323B2 (en) Method and system for extracting information from unstructured text using symbolic machine learning
US10698977B1 (en) System and methods for processing fuzzy expressions in search engines and for information extraction
Quan et al. Construction of a blog emotion corpus for Chinese emotional expression analysis
US11768884B2 (en) Training and applying structured data extraction models
US9323741B2 (en) System and method for searching functions having symbols
Srihari et al. Infoxtract: A customizable intermediate level information extraction engine
US8364470B2 (en) Text analysis method for finding acronyms
US20150227505A1 (en) Word meaning relationship extraction device
JP4778474B2 (en) Question answering apparatus, question answering method, question answering program, and recording medium recording the program
EP1616270A1 (en) Method for sentence structure analysis based on mobile configuration concept and method for natural language search using of it
Moldovan et al. An interactive tool for the rapid development of knowledge bases
Hussein Arabic document similarity analysis using n-grams and singular value decomposition
Hussein Visualizing document similarity using n-grams and latent semantic analysis
Williams et al. Understanding and inferring units in spreadsheets
EP1290574B1 (en) System and method for matching a textual input to a lexical knowledge base and for utilizing results of that match
Bhat Morpheme segmentation for kannada standing on the shoulder of giants
Thalib et al. A review on question analysis, document retrieval and answer extraction method in question answering system
Khamphakdee et al. A Framework for Constructing Thai Sentiment Corpus using the Cosine Similarity Technique
Villavicencio et al. Discovering multiword expressions
Litvak et al. Multilingual Text Analysis: Challenges, Models, and Approaches
Greenbacker et al. Improving the accessibility of line graphs in multimodal documents
CN114722224A (en) Image-text cross-modal retrieval method based on joint features
Vetriselvi et al. Latent Semantic Based Fuzzy Kernel Support Vector Machine for Automatic Content Summarization.
Wimalasuriya Automatic text summarization for sinhala
Younas et al. An Artificial Intelligence Approach for Word Semantic Similarity Measure of Hindi Language.

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JOHNSON, DAVID E.;OLES, FRANK J.;REEL/FRAME:015584/0447

Effective date: 20041101

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION