US20050065948A1 - Test generator for database management systems providing tight joins - Google Patents

Test generator for database management systems providing tight joins Download PDF

Info

Publication number
US20050065948A1
US20050065948A1 US10/978,965 US97896504A US2005065948A1 US 20050065948 A1 US20050065948 A1 US 20050065948A1 US 97896504 A US97896504 A US 97896504A US 2005065948 A1 US2005065948 A1 US 2005065948A1
Authority
US
United States
Prior art keywords
block
statement
statements
database
test
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/978,965
Inventor
Donald Slutz
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Microsoft Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/078,837 external-priority patent/US6138112A/en
Application filed by Microsoft Corp filed Critical Microsoft Corp
Priority to US10/978,965 priority Critical patent/US20050065948A1/en
Publication of US20050065948A1 publication Critical patent/US20050065948A1/en
Assigned to MICROSOFT TECHNOLOGY LICENSING, LLC reassignment MICROSOFT TECHNOLOGY LICENSING, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MICROSOFT CORPORATION
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/36Preventing errors by testing or debugging software
    • G06F11/3668Software testing
    • G06F11/3672Test management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/21Design, administration or maintenance of databases
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/242Query formulation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99932Access augmentation or optimizing
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99933Query processing, i.e. searching

Definitions

  • the present invention relates to electronic data processing, and more specifically concerns automated testing of database management systems using statements having a tight join of a plurality of tables.
  • DBMS Relational database management systems
  • SQL Structured Query Language
  • SQL Structured Query Language
  • Testing these systems is recognized throughout the industry as a technical challenge of the first magnitude. SQL and similar database-system interpreters are highly complex. For example, they offer sophisticated optimization techniques and execution planning for queries input on the fly; opportunities for arcane design problems are ubiquitous. At the same time, the state space to be tested is gigantic. For a one-gigabyte database, the possible combinations of database configuration and SQL statement to be executed exceeds 10 2,000,000,000 .
  • Libraries of test scripts for relational database systems typically contain thousands or tens of thousands of sample statements which are applied to a test database for comparison of their results with known correct data.
  • Existing libraries are known to be inadequate; most commercial database systems produce a constant and substantial stream of reported bugs.
  • the amount of work required to generate larger libraries quickly becomes prohibitive.
  • test-case generator may choose a random mix of hand-generated fixed scripts. Choosing random parameter values in fixed scripts increases the effective number of test cases. These methods still require painstaking human composition and verification of long, multilevel queries. Automated generation of very simple queries considerably speeds up the generation of test cases, but eliminates the more complex test cases where subtle errors lurk.
  • test systems are effectively limited to a fixed database, or to simple variations on fixed data.
  • the test system In order to construct statements that actually execute properly against the target database, the test system must be internally coded to produce only those statements that match the semantics of the database, the names of the database tables and their columns, the particular data types of each column, and so forth.
  • testing on only one set of data obviously restricts the range of the tests that can be performed and thus the errors that will be uncovered.
  • the use of fixed data for many tests does not permit slanting test runs toward certain kinds of applications, or focusing on the kinds of data or database structures that have been found to produce errors.
  • test statements although desirable for teasing out subtle errors, works against the isolation of bugs which cause those errors. Short, simple statements that produce errors are much more useful for tracking the errors down to particular parts of the DBMS under test.
  • test SQL statement refers to data in multiple SQL tables
  • the set of result rows often includes all combinations of data from rows in the tables in the From clause. This is called a Cartesian product of the tables.
  • the number of output rows is the product of the sizes of the queried tables, which can impose artificial limits in the sizes of tables in the test database.
  • a second shortcoming results from the use of randomly generated character string constants.
  • Some SQL database systems support special searching of large text objects. For example, the text objects might be newspaper articles and the search predicate would specify articles that contained two particular phrases “near” each other. Near could mean anywhere in the same paragraph.
  • the text predicates concentrate on finding words or strings that are in the text. Generating character strings constants with randomly selected characters almost always yields tokens that are not found in the text string. This reduces the effectiveness of random testing since some code paths will not be tested much.
  • the present invention speeds up the generation of database test cases by orders of magnitude.
  • a typical generator running on a personal computer having a single 200 MHz microprocessor outputs 700 SQL statements per second, about a million times faster than a human.
  • the queries are complex and can have multiple nested levels. They have valid semantics as well as valid syntax; that is, they will run correctly on a bug-free database system, using whatever sample database is selected for a test run.
  • the statistical and other features of the test cases are configurable.
  • a test operator may choose the syntactic elements selectable in queries or other statements, the frequency of their use, and parameters such as the maximum subquery depth.
  • the invention achieves these and other objectives by reading configuration data containing a set of test parameters, reading the schema of an arbitrary database, then constructing a number of test statements that are syntactically correct for the DBMS being tested, that are semantically compatible with the target database, and that have content and characteristics pursuant to the configuration data.
  • One or more DBMSs under test execute the statements and return result data. Execution errors are detected, as well as result-data differences. Error-producing statements can be converted into greatly simplified statements that provoke the same error, in order to facilitate fault isolation.
  • Generated SQL statements include predicates that are tightly joined in order to avoid results sets that comprise the Cartesian product of the data in the tables.
  • a From list contains N tables.
  • a list of N sets of table names is then created. Initially, each table set contains one table name from the From list. With each iteration, pairs of table sets are uniformly selected, and a table is uniformly selected from each table set.
  • a column from each selected table is chosen and a predicate equating, or otherwise relating, the two columns is ANDed into the Where clause.
  • the two selected table sets are merged into one, and the two selected table sets are then removed. The process iterates until a single table set remains.
  • a further aspect of the system is that text strings to be included in predicates can be selected from a dictionary.
  • the dictionary is created by sampling the text columns in the target database and extracting a random collection of actual words to place in the dictionary. Words from the dictionary can be randomly interspersed with randomly generated words to form argument values for the full text predicates.
  • the fraction of dictionary words used compared to randomly generated words can be a configuration parameter of the automated SQL testing tool.
  • a separate dictionary is built for each text column in the database.
  • the number of words, to be placed in the dictionary can be a configuration parameter (specified either as a constant or as a percentage of the total number of bytes in the text column).
  • FIG. 2 is a block diagram of an illustrative database system to be tested by the invention.
  • FIG. 3 is a flow diagram for a test generator according to the invention.
  • FIG. 4 comprising FIGS. 4A-4F , shows a configuration file used in the process of FIG. 3 .
  • FIG. 5 shows an example of a syntax diagram used for constructing a statement in FIG. 3 .
  • FIG. 6 is a flow diagram detailing the construction of a test statement according to FIG. 3 .
  • FIG. 7 illustrates a state-information file for use in conjunction with an implementation of FIG. 6 .
  • FIG. 8 shows an example of a parse tree for a statement constructed by the process of FIG. 6 .
  • FIG. 9 comprising FIGS. 9A-9B , illustrates a more typically complex test statement produced by the process of FIG. 6 .
  • FIG. 10 is a flow diagram for simplifying a faulty statement according to FIG. 3 .
  • FIG. 11 shows a simplified version of the statement of FIG. 8 produced by the process of FIG. 9 .
  • FIG. 12 is a flow diagram for providing a tight join for test statements produced by the process of FIG. 6 .
  • FIG. 13 provides an exemplary application of the process illustrated in FIG. 12 .
  • FIG. 14 is a flow diagram for generating text strings for statements produced by the process of FIG. 6 .
  • FIG. 1 is a high-level block diagram of a conventional client/server computer system 100 .
  • Network wiring 110 interconnects a number of personal computers (PCs) 120 to a server 130 via network adapters 121 and 131 .
  • Server 130 includes a storage subsystem 132 for holding the large amounts of data in typical enterprise databases.
  • Other system architectures are also suitable environments for the invention; for example, units 120 may be terminals connected to a mainframe or midrange computer 130 , or unit 130 may itself comprise a PC coupled to PCs 120 in a peer-to-peer network.
  • the entire system 100 may comprise a single PC acting as both client and server.
  • file storage may be distributed among a number of different machines.
  • FIG. 1 includes a schematic representations of an external storage medium 133 which may store client and server software for distribution and downloading to clients, and another medium 134 , such as a diskette, for offline storage of database tables.
  • Medium 134 can also store instructions and data for the test program of the present invention; the test program can be executed in one or more of the clients 120 , or even in server 130 .
  • FIG. 2 is a block diagram of a typical conventional client/server database management system 200 capable of operating in system 100 , FIG. 2 .
  • a client application program 210 executes within each PC 120 , under a PC operating system 220 such as a version of the Microsoft Windows operating system, including Windows 95®, Windows 98®, Windows Me®, Windows NT®, or Windows 2000®.
  • client application 210 contains a facility 211 for accepting database queries from a user at a PC 120 .
  • other application programs 230 executing in some of the PCs 120 may present queries to DBMS client 210 , via predefined host-language application-program interfaces (APIs) 231 .
  • APIs application-program interfaces
  • One of these programs can be the test program of the invention.
  • a DBMS server application 240 such as Microsoft SQL Server, executes under a server operating system 250 such as Microsoft NT.
  • DBMS program 240 provides services for creating, maintaining, and modifying a number of relational databases, exemplified by database 260 .
  • Program 240 may employ the file-system services 251 of operating system 250 , or may provide its own file system.
  • Operating system 250 could execute a separate instance of the entire DBMS application for each request from a client 210 .
  • program 240 gives each client connection a separate thread 242 in the DBMS kernel. Further, this thread is a native operating-system thread, which carries with it all the Windows NT mechanisms for process memory protection, better access to storage devices, and so forth.
  • Search engine 241 processes queries from individual clients 210 upon tables 261 of a database 260 , as described more fully below. It also enforces database integrity with conventional facilities for record locking, atomic transactions, etc.
  • the interface language between query facility 211 and search engine 241 is Transact-SQL, which provides much of the function of the standard ANSI SQL, 89 and ANSI SQL 92 languages, plus extensions for providing greater flexibility and programmability.
  • the test program may execute query-language statements upon multiple systems, either concurrently or sequentially.
  • the reference numeral 240 ′ indicates a second DBMS, which manages the same database 260 as the first system 240 (either at different times or, for some DBMSs, concurrently), or a different database, indicated as 260 ′.
  • FIG. 3 shows the overall flow of a test generator 300 for producing massive numbers of statements for testing DBMS applications 240 according to the invention.
  • Test program 300 can reside in one of the PC clients 120 , FIG. 1 , where it may simulate a user, alternatively, it can execute within server 130 , or at any other convenient location.
  • Testing a DBMS in this context concerns checking the correctness of its implementation of the language by which it interfaces between databases and users of the databases.
  • These languages are generally referred to as query languages, even though their statements perform in any functions other than querying data in tables; for example, statements can create and modify data tables, insert and modify table data, define transactions for committing data, manage database security, and compile performance statistics.
  • query is usually used as a synecdoche for any statement in an interface language.
  • This embodiment assumes the use of SQL (Structured Query Language), which is actually a family of more or less standardized dialects in common use. Any of these dialects will serve the present purposes.
  • Other query languages such as QBE (Query by Example”) can be substituted easily. Every query language has an explicit syntax specification that defines how to construct valid statements in the language.
  • One aspect of testing the DBMS involves determining whether a statement that is well-formed according to the syntax of the query language does in fact execute without error in the tested DBMS (or, more generally, whether the DBMS reports the proper error status of any statement given to it).
  • Another aspect involves verifying that the results (or lack of results) that a statement produces in that test database are correct. The present embodiment can perform both of these tests.
  • Preliminary blocks 310 begin the entire testing process.
  • Block 311 reads in a configuration file 400 containing a set of parameters for the test procedure.
  • One of the parameters specifies the name of a database 260 to provide the data tables 261 . That is, the process is not limited to one or more fixed databases for testing a DBMS, but can employ arbitrary, user-selected target databases.
  • Another parameter is able to specify the name of a different DBMS as a verification DBMS for checking the correctness of returned results.
  • Other parameters of the configuration file will be described below.
  • Block 312 connects test program 300 to test DBMS 240 using the standard Open Database Connection (ODBC) protocol or any other conventional method.
  • Block 313 determines the schema of the test database 240 .
  • the schema of a database is the organization of its data. Schemata for relational databases include the names of all data tables, the names, positions, and data types of table columns, restrictions on null or duplicate data values, and so forth.
  • the test-database schema may be derived from any convenient source; it could, for example, be stored in the configuration file.
  • Block 314 symbolizes the syntax specification of DBMS 240 . In this implementation, the syntax is built into the code that constructs statements, rather than read in from an external source.
  • Block 320 constructs each statement to be used in the testing process as a parse tree by following the syntax specification of block 313 .
  • block 320 inserts a syntactic element which comports with the test-database schema read by block 314 .
  • the selection of one of a number of grammatically correct elements is made by a random roll from a seed, guided by probability parameters contained in the configuration file from block 311 .
  • the configuration file can, in fact, specify that only certain features, certain syntactical constructs, or certain parts of the database be included in the statements.
  • block 320 produces statements using a pseudo-random number generator 321 , so that the same configuration settings, the same schema, and the same starting seed cause it to produce the same statement deterministically; this allows regression testing of the DBMS. Although one could save the sequence of generated statements for regression testing, the advantage here is that saving only the starting seed can reproduce the entire sequence later.
  • block 320 follows a set of internal rules that further permit or constrain certain choices at certain points in the parse tree.
  • Block 330 causes test DBMS 240 to execute the statement against database 260 . If the configuration file had specified verification testing, then block 330 ′ causes verification DBMS 240 ′ to execute the same statement against the database. Each execution may or may not produce a set of data (for a query) or other results from the database. Verify block 331 compares the result sets produced by multiple executions, and produces a verify-error indication if they are not the same. Although the results could be compared in detail, a simpler check usually suffices; for example, block 331 can compare the number of affected rows for a data-modification statement.
  • Block 331 can count the rows in the result sets of a query statement, then generate and compare checksums over the column values in all the rows; this avoids having to sort the result data and compare it one unit at a time.
  • the configuration file can specify a round-off tolerance for numeric fields. Date/time fields are more problematic, because the same value might assume different valid forms; a configuration parameter can specify a common format, or can specify that these fields are not to participate in comparisons.
  • Block 330 also produces an operational error indication when test DBMS 260 fails to execute the statement properly. Besides execution errors (usually including query-language compiler errors) reported by the DBMS, operational errors include lost connections to the database, deadlocks, and system crashes. If a DBMS connection is lost, program 300 attempts to reconnect; if this attempt fails, the program aborts the test run. A system crash requires an automated rapid restart of the server in order to continue the run.
  • Block 332 logs each executed statement, its result data, and any error indications, along with the seed value for each statement. If verification is not being done for the test, logging this data allows the test to be run again at a later time in a pseudo-verification mode with the same DBMS 240 , using logged data from the previous run for comparison. This mode is useful for catching suspected intermittent errors.
  • Blocks 340 process data and operational errors resulting from the execution of a statement. If no error occurs, block 341 passes control directly to loop control 350 .
  • the extremely long test runs generated by program 300 can result in huge error files of ‘uninteresting’ errors.
  • An uninteresting error is one that is expected to naturally occur in a randomly generated statement. For example, divide-by-zero and overflow errors are uninteresting. Therefore, block 342 filters some of these from the log, or merely counts their occurrences. Additionally, some errors tend to occur in large numbers when they occur at all. For this situation, a configuration-file parameter specifies certain error codes to count or to ignore entirely.
  • Block 343 simplifies statements which have produced errors. Debugging a complex non-procedural program such as a DBMS is greatly facilitated by modifying failed statements to produce simpler versions that still produce the same error. Simplification is particularly advantageous for the long, complex statements produced by program 300 . Simplification proceeds by sequentially removing as many elements of the statements as possible while preserving the same error indication. The simplified statement is usually not equivalent to the original statement which means it would not return the same result set had no error in the system occurred. Although one bug might cause a certain error indication in the original statement and a different bug in the simplified one; this has been found to be extremely unlikely. Block 344 records the simplified statement and keys it to the original.
  • Block 350 passes control back to block 320 to generate another statement as long as the test run has not completed. Completion conditions are read from the configuration file. Any convenient measure, such as number of statements, number of errors, and run time, are possible.
  • Block 351 draws up a report of the entire test run.
  • the report can include conventional items such as error listings and statistics concerning the different types of features included in the test statements, or reduced summaries of such items.
  • the report also optionally lists the schema of the test database and the current configuration-file settings. For the pseudo-verify mode described above, it is also useful to include machine readable information on each statement, its result data, and error indications.
  • Program 300 can then run these same statements against the test database on the same test DBMS at a later time, and compare the two runs. Also, a conventional utility program can combine and summarize the results of multiple test runs carried out concurrently on different data processors.
  • This embodiment of program 300 generates and executes an entire suite of test statements in a single run. It is entirely possible, of course, to generate all the test statements in a batch, then to run those statements from a file at a later time. For example, program 300 could execute blocks 310 , then loop on block 320 until all statements have been completed. Then a later run could loop on blocks 330 - 340 , and print the report of block 351 .
  • FIG. 4 shows a portion of a configuration file 400 containing about a hundred parameters that program 300 employs to control a test run.
  • the format is similar to that of a conventional .INI file for specifying parameters and data to a program. Each line names a parameter and gives it value; comment lines begin with asterisks or dashes.
  • a Program section specifies global parameters controlling the overall operation of a test run, including the database schema of block 313 .
  • the parameter fSimplifySQL determines whether or not to auto-simplify faulty statements.
  • a Console section controls data display for real-time tracking of a test run.
  • An SQL file section configures the report file of block 351 .
  • a Limits section specifies bounds on database size, overall statement size, subquery depth, number, and so forth.
  • An Allowed section controls the type and content of statements generated by block 320 .
  • Many of the parameters occur in pairs naming a maximum and a frequency for statement features or constructs.
  • a group of parameters near the end of this section specify the mix of different statement types, such as UPDATE, INSERT, DELETE, and SELECT.
  • transactional integrity is an important aspect of DBMS operation, other parameters control the use of multi-statement transactions and methods, COMMIT and ROLLBACK, to end the transactions.
  • Configuration file 400 not only permits the testing of different systems and databases, but also greatly facilitates the isolation of system bugs. That is, particular features and statement characteristics that produce problems can be emphasized merely by changing a few parameters in the file.
  • Program 300 builds statements in response to parameters from the configuration file, to database structure from the schema, and to the syntax of the query language.
  • Query-language syntax is usually written as a set of productions in a stylized script or diagram format such as Backus normal form.
  • a syntactically valid statement is conceptually constructed by expanding the productions, choosing only one of the allowable paths at every node where the syntax permits alternatives.
  • individual procedures build major productions such as WHERE and GROUP BY; the smaller productions, such as picking operators in an expression, employ simple case statements.
  • FIG. 5 shows a partial set of productions 500 for a SELECT statement in a simplified dialect of SQL.
  • the name of each production is printed above its diagram. Rectangular boxes indicate other productions. Rounds and ovals indicate terminal syntactic elements. Small diamonds indicate points at which alternative paths exist.
  • Block 511 requires that a SELECT statement begin with the word “SELECT”, which can be optionally followed by a qualifier word “ALL” or “DISTINCT” at point 512 . Then a Select List Entry 520 can be repeated zero or more times at node 513 , separated by commas.
  • Production 520 for a Select List Entry comprises either one or more Column Name syntactic elements 521 separated by Operator elements 522 at choice point 523 .
  • point 524 can specify a Function name 525 followed by a parenthesized Column Name 521 .
  • Production 540 reveals that one of the elements of this predicate is a Select Statement production 510 . That is, an entire SELECT statement can be nested inside another SELECT statement; such a nested occurrence is referred to as a subquery, and its containing query or subquery is called an outer query. A column C belonging to a table in an outer query can be referenced in a subquery. The reference to C in the subquery is called a correlated reference.
  • FIG. 6 shows a process 600 for constructing a single test statement according to block 320 , FIG. 3 , by following syntax diagrams stochastically.
  • routine 600 does not produce any explicit representation of a parse tree.
  • file term “roll” means to obtain a number from random number generator 321 and use the number to make a choice among alternatives allowed by certain constraints, with probabilities which might be specified by rules, settings, or other sources; it includes situations where the constraints, etc., might allow only a single choice.
  • Block 610 rolls to determine the type of statement to be generated.
  • the alternatives include SELECT, UPDATE, INSERT, etc.
  • Configuration file 400 lists the desired probabilities of each statement type.
  • Block 611 initializes information described below that is carried along from node to node during the process.
  • Block 620 follows the syntax diagram of the selected statement. That is, at each activation, it chooses the next syntactical element or set of alternative elements in the diagram as the current node.
  • block 620 would first choose element 511 , then the alternatives presented at point 512 , then, depending upon a roll outcome, neither or one of the elements 521 and 525 , etc., then onward through diagram 520 .
  • a roll at point 513 sends block 620 through element 520 again, or on to element 514 .
  • Some elements might be visited recursively and/or repetitively as block 620 travels upward and downward through the diagrams 500 .
  • Block 621 outputs a representation of the statement to program 300 as the statement is generated. If the statement will be executed, block 621 also places a copy in a buffer.
  • process 600 works on one syntactic element at a time to build one node of the parse tree. At certain places, however, it is necessary to gather preliminary information and to make certain preliminary choices involving elements other than those at the current node in the diagram. For example, the names of the tables to be used in a SELECT query need not be determined syntactically until block 620 reaches element 515 in FIG. 5 . However, the choice of columns at previously encountered element 520 could prow3ke inconsistencies with certain constraints, such as parameter values in configuration file 400 or the schema of the database. Therefore, block 630 obtains any preliminary information specified for the current node, block 631 rolls any choices specified by the information, and block 632 records choices that might be useful or might affect choices further along the tree.
  • block 630 retrieves the list of tables from the schema and the maximum number of tables in a join parameter from configuration file 400 .
  • Block 631 chooses a list of particular tables randomly within the parameter limit, and reads the names and data types of all columns belonging to these tables.
  • Block 632 records the names of these columns, because this choice affects the alternatives available to the subsequent element 515 , and also to other elements, such as the Column List of GROUP BY clause 517 .
  • block 632 also records the data types of these columns in order to avoid having to look them up again while processing subsequent elements involving functions of those columns.
  • Block 640 gets information pertaining to the current node that can affect the alternatives available at that node. State information such as the identity of the present element in the syntax diagram obviously affects this choice. Certain parameter values in configuration file 400 might be relevant to the current node; for example, the maximum subquery depth might bar the choice of element 540 at point 531 in FIG. 5 , because this element always generates a subquery at element block 510 . Some of the preliminary choices might affect the current choices; this includes new information recorded by block 632 for the current node and for any higher node that may affect the current choice. For instance, the preliminary choice of less than all tables, made at element 515 , precludes later nodes from choosing columns in other tables of the database schema.
  • Block 641 presents a cumulative history of previous choices made for any previous nodes in the parse tree.
  • previous Choices can affect subsequent alternatives at the current node.
  • the data type of an expression is chosen before the expression is generated so the data type information is passed down the tree. This is true even when the alternatives might be syntactically correct; for instance, an ORDER BY entry can be an integer to index the select list but the integer value can not exceed the number of elements in the Select List.
  • blocks 640 and 641 The difference between blocks 640 and 641 is that the first restricts alternatives at the current node based upon determinations made for nodes higher (toward the root) in the constructed parse tree, while block 641 carries choices made at lower (toward the leaves) nodes upward to restrict choices when the current node is higher in the tree and choices made at sibling nodes in other subtrees of the parse tree. Stated another way, block 641 concerns all the node choices made up to the current node that are not in the path from the current node to the root. Carrying the results of choices to higher nodes is one of the most important factors in constructing semantically correct test statements for arbitrary databases according to the invention. Without a facility for carrying previous choices upward in the tree, process 600 would have to cleave closely to a predefined fixed database in order to construct statements consistent with the semantics of the database.
  • Block 642 makes available a set of rules written for particular nodes. These rules are of two types.
  • Mandatory rules require the performance of certain actions. For example, some points in a syntax diagram require that a column reference be generated; thus any potential alternatives at the current node that do not result in a column reference must be excluded.
  • Optional rules permit certain actions. Such a rule might allow the generation of a column reference to a correlated column in a subquery at the current node. Rules can be hard-coded into a program that executes process 600 , stored in an external file, or implemented in any other convenient manner.
  • Block 650 assembles a list of alternatives available at the current node. These alternatives arise from the set of elements specified at the current node of the syntax diagram, and are potentially modified or restricted by some or all of the information obtained in blocks 640 - 642 .
  • Block 651 rolls the assembled alternatives to produce a random choice of a syntactic element for the current node. (As a shorthand notation, the term ‘roll’ means to choose among a set of alternatives in accordance with a random or pseudorandom number; the probabilities of each alternative can be equal or weighted.) This can include a terminal value; e.g., a roll might produce the numeric value 10000 where the current node requires an item of numeric data.
  • Some of the alternatives might have probabilities or ranges specified by weighting parameters in the configuration file, as mentioned.
  • the roll outcome could be derived from a non-predictable random source, such as hashing the time of day, using a pseudo-random roll allows a seed value recorded in the configuration file to reproduce the action of block 651 deterministically if desired for regression testing or other purposes.
  • block 651 or blocks 640 - 650 , can be entirely bypassed to save time.
  • Block 652 records the choices made in the history obtained by block 641 .
  • Block 653 updates the information obtained by block 640 , based upon the current-node choices.
  • Block 654 updates the statement according to the choices made in block 651 . Normally, a choice at each current node adds a term to the statement's parse tree. However, it is sometimes advantageous to delay this action. For example, column references are constructed for a GROUP BY clause 517 , FIG. 5 , but later choices can generate expressions and intermix them randomly with the column references before outputting the clause as part of a SELECT statement or subquery.
  • process 600 can be implemented as a number of individual functions called recursively on a parse tree.
  • a C function GenSelect produces a single SQL statement 510 at any level of a parse tree.
  • this function first rolls to determine the list of tables in the FROM clause, within the configured limit. If this query must return a particular data type, it checks to ensure that at least one listed table contains a column of that type. Next, the function builds a list of data types for all columns in the selected tables. Another roll determines how many elements the select element 520 will contain. The next roll determines whether the query will include a DISTINCT clause, and/or will use the DISTINCT qualifier in aggregate functions.
  • a roll includes or omits a GROUP BY clause, and, in the former case, rolls to build its column references—but does not yet output the clause.
  • the next step is to output the “SELECT” keyword to the tree, and start following syntax diagram 500 .
  • a sequence of rolls generates expressions for syntactic element 520 which are then output.
  • the keyword “FROM” and the list of tables in the FROM clause is then output to the parse tree.
  • a sequence of rolls generates their ON-clause predicates.
  • a roll includes or omits a WHERE clause 516 , and performs further rolls to build it. If a predicate involves a subquery, GenSelect ( ) produces certain rules (e.g., return no more than one row of data), and calls itself recursively to build the subquery.
  • a further roll possibly includes and builds a HAVING clause, building any subqueries and their rules as above.
  • a roll possibly includes and builds an ORDER BY clause, then rolls to intermix expressions and index references to entries in the Select List.
  • GenExpr Another function, GenExpr ( ), builds expressions for GenSelect ( ) and for other functions. It outputs a constant or a column reference, or calls itself recursively to generate operators or functions that take expressions as arguments. While the design of such a function is conventional, this particular expression builder illustrates the use of rules according to the invention.
  • a call to GenExpr( ) includes two parameters. TableInfo holds the names and data types of table columns, as described above.
  • ExprInfo contains the explicit rules that the expression must follow, and includes state information that influences the function's choices for expression terms.
  • FIG. 7 lists the parameters of the ExprInfo rules and their meanings. Element names beginning with ‘f’ are binary flags.
  • the rule element DataType e.g., is a mandatory rule enforcing a particular data type resulting from the evaluation of the expression, while fLocalFunctions is an optional rule allowing aggregate functions such as AVG( ) for any column contained in a table in the current FROM clause.
  • FIG. 8 shows an example of a parse tree 800 for a very short SELECT statement, as generated by process 600 .
  • the root of the tree is the SQL keyword “SELECT”, 810 .
  • the next level of the tree has branches to three nodes.
  • block 820 of the select list 520 FIG. 5
  • process 600 proleptically produces a pool of possible table and column names, as described above.
  • Block 820 specifies that the name column of the database's employee table is the first item in the list.
  • the second item in the Select List is an expression, the sum 821 of columns named salary 822 and commission 823 .
  • Nodes 822 and 823 are leaf nodes; that is, their contents specify terminal elements in the syntax of a SELECT statement.
  • the second high-level tree branch leads to node 830 , the keyword “FROM” for table list 515 .
  • the only table included in this statement is employee, at leaf node 831 . Again, process 600 had chosen this table, before actually outputting nodes 820 - 823 ; therefore, the column nodes in 820 - 823 were restricted to be columns of the employee table.
  • the third high-level branch specifies a WHERE clause at node 840 .
  • Nodes 841 - 847 constitute a predicate generated by a call to a function GenPred( ).
  • Node 841 defines this predicate as a logical AND of two subpredicates.
  • Nodes 842 - 844 generated by a recursive call to GenPred( ), define the first subexpression as the condition that the value of the salary column exceed the numeric constant 10000 .
  • Nodes 843 and 844 are generated by calls to GenExpr( ).
  • the second sub predicate, nodes 845 - 847 specifies rows of the employee table where the department column contains the string constant “sales”.
  • Process 600 builds statements in the form of parse trees such as 800 .
  • program 300 normally outputs test statements in the same form that a user or another program would input them to DBMS 240 , FIG. 2 . Therefore, output block 621 in FIG. 6 outputs each statement as an equivalent character-based representation.
  • Parse tree 800 for example, becomes the statement:
  • FIG. 9 shows a more typically complex SQL statement generated by process 600 for a publishing-company database.
  • Statement 900 nests subqueries up to five deep, and the inner queries reference correlated columns in the outer queries.
  • the statement syntactically follows the DBMS query language, and the requested data comports semantically with the rules of SQL and with the schema of the target database.
  • some of the more prolix constructs, such as adjacent minus signs and redundant parentheses frequently provoke design errors that more usual statements overlook.
  • the maximum join size and the maximum subquery depth should be kept low in order to avoid excessive run times and system limits.
  • ten different clients 120 FIG. 1 , concurrently executed 25,000 SQL statements equally distributed over SELECT, INSERT, UPDATE, and DELETE types.
  • Expected errors mostly deadlocks
  • Unexpected errors having two different error codes occurred in a total of eighteen statements.
  • FIG. 10 shows a process 1000 for simplifying an error statement (that is, one which has produced an error indicating a fault in DBMS 240 ) according to block 342 , FIG. 3 .
  • an error statement that is, one which has produced an error indicating a fault in DBMS 240
  • the query compiler of DBMS 240 attempts to reduce processing time and memory usage with complex access plans involving many steps. Debugging the complex plans step by step is extremely tedious and difficult. Therefore, finding a simple statement having the same error as a much more complex statement aids greatly in isolating the cause of the error.
  • Block 1010 first reads in the error statement detected by block 341 in the format of a parse tree as shown in FIG. 8 .
  • Block 1011 records an indication denoting the particular error that occurred; usually, this is a code produced by search engine 241 , although it can also be a signal indicating a connection loss, or some other indication.
  • Block 1012 initializes process 1000 to the root node of the parse tree (e.g., node 810 , FIG. 8 ).
  • Block 1020 walks the parse tree of the error statement, using any conventional method.
  • Block 1020 chooses certain nodes whose subtree is a candidate for removal.
  • Block 1030 removes the entire subtree of the current node. (For some cases, such as expressions involving two operands such as ‘A+B’, block 1030 first removes subtree A and also the ‘+’. Then it removes subtree B and also the ‘+’.)
  • Block 1031 then reexecutes the truncated statement on the search engine. If the search engine returns the same error indication, block 1032 passes control to block 1020 to determine whether this truncated statement can be further simplified by removing additional subtrees.
  • block 1033 restores the subtree removed by block 1030 before passing control back to block 1020 .
  • block 1021 translates the truncated tree into the form of a simplified statement and outputs it to block 344 , FIG. 3 .
  • block 1020 first visits the branch to node 820 , to remove the subtree including nodes 820 - 823 , then the branch to node 821 , to remove nodes 822 - 823 , then to remove 822 alone, then to 823 alone.
  • the branch to node 830 is removed, and that subtree is followed to its leaf 831 .
  • the branch to node 840 is removed, and its subtree is visited.
  • block 1020 need not attempt to remove each possible subtree of the parse tree. For example, removing the entire subtree 820 - 823 would result in a simplified statement having incorrect syntax, because the entire select list would be missing from the SELECT statement. Block 1032 would then reject the resulting syntax-error code, causing block 1033 to restore the subtree. Removing the subtree containing nodes 821 - 823 , however, leaves the statement in a syntactically correct form. ( FIG. 8 has been simplified for this short example. For more complex statements, removing a subtree does not remove a suffix of the Select List. Select List entries are not removed, but rather only elements of their expressions. In expressions with two operands, the operator is removed along with the subtree of one of the operands.)
  • a more intelligent block 1020 understands at least some of the formalities of the target query language, and passes over branches which are known to be syntactically required or to possess other properties rendering them unsuitable for deletion.
  • a number of heuristic optimizations have been incorporated into the statement simplifier. For example, it performs trial deletes from the top of the tree down; this alone eliminates the processing of most of the subtrees. The simplifier does not try all possible deletions. For example, it does not lop items from the Select List or from the ORDER BY list. It does not always plumb the deepest level. It never back-tracks. Once it tries a delete and finds it doesn't work, it never tries the same delete again, although there are cases where other successful deletes enable a previously tried one.
  • FIG. 11 shows a simplified statement generated from the statement of FIG. 8 by process 1000 .
  • Such drastic parings of the parse tree are not uncommon. Modifying elements of the simplified statement for further fault isolation is now feasible, whereas modifying portions of the full statement of FIG. 4 would be a daunting task.
  • an SQL join is the combining of two (or more) SQL tables, often with a predicate that relates rows in different SQL tables. For example, suppose SQL table Employee has columns Name, DeptNumber, and Salary and contains 1000 rows. Also suppose that SQL table Department has columns DeptNumber, DeptName, Manager, and Location and contains 100 rows.
  • the SQL query Q1: Select Employee.Name, Employee.Salary, Department.Location From Employee, Department (Q1) is an example of a join of the Employee and Department tables.
  • the result of executing the above query is a table with 3 columns and 100,000 rows representing all combinations (1000 ⁇ 100) of rows in both tables.
  • the result table is the size of the Cartesian product table. In joins of tables where each table has many rows, the Cartesian product can grow quite large, and it can take a large amount of time to return the complete result set.
  • the query in Q2 filters the results to exclude those employees with a DeptNumber value that does not match a value in Department.
  • FIG. 12 is a flow diagram of a process 1200 for providing a tight join for test statements produced by the systems and processes described above in reference to FIGS. 1-11 .
  • the process begins at block 1202 , where a SQL Select statement is generated as described above.
  • each table set initially contains one table in the From list.
  • the I-th table set will contain the I-th table in the From list.
  • a pair of table set TSj and TSk are selected from the plurality of table sets.
  • the table set pairs are selected uniformly, that is, randomly.
  • other selection algorithms can be used.
  • the least recently used table set could be selected.
  • the invention is not limited to any particular mechanism for selecting a table set pair.
  • a table T 1 and Tm are selected from table sets TSj and TSk respectively.
  • the table sets will contain more and more tables.
  • the selection of a particular table from a table set is uniform.
  • the tables are selected using other algorithms known to those of skill in the art.
  • columns Cx and Cy are selected from tables T 1 and Tm respectively.
  • the columns are selected uniformly.
  • the columns are selected according to whether they have a compatible data type such that they can be logically compared with little or no type conversion (i.e. conversion of a numeric to a character string or vice versa).
  • a list of column pairs that have the same or similar column names is created.
  • a column pair can be considered similar if the pair has a common prefix or suffix, are phonetically similar, or exhibit a common pattern.
  • a column pair is then chosen uniformly from the list of similar column pairs. Selection of similar column names is desirable, because it is often the case that similar column names represent foreign keys relating on table to another, and thus emulates the manner in which select statements will be used by actual customers.
  • a predicate is formed using the two column names.
  • the main implementation is for the predicate to perform a logical comparison to determine if the data value for column Cx is equal to that for column Cy, but other types of comparisons are possible that preclude Cartesian products.
  • the predicate is “ANDed” into the existing Where clause for the SQL statement.
  • the tables in table sets TSj and TSk are merged into a single table set, and TSj and TSk are deleted.
  • a decision block 1216 determines if more than one table set remains after the merging. If so, the process proceeds to block 1206 to repeat the predicate creation process. If not, the Where clause is complete and the process ends.
  • the Where clause that is generated represents a tight join of the tables in the From clause, and restricts the results set such that a Cartesian product of the rows in the tables is avoided.
  • FIG. 13 provides an exemplary scenario of the application of the process described in FIG. 12 , in which a SQL Select statement is transformed to include a Tight join.
  • the original query 1302 . 0 is shown in along with an initial five table sets 1304 . 0 each containing one table from the From list.
  • Table sets ⁇ T3 ⁇ and ⁇ T5 ⁇ are merged into a single table set ⁇ T3,T5 ⁇ resulting in table set 1304 . 1 .
  • Each iteration of process 1200 reduces the number of table sets 1304 . 2 - 1304 . 4 by one until only one table set 1304 . 4 remains.
  • the SQL statements that are created in each iteration are represented in 1302 . 2 - 1302 . 4 .
  • the examples presented above illustrate a top-level Select statement with a From clause consisting of simple tables.
  • the invention is not so limited.
  • the Tight Join process can also applied in exactly the same way to other scenarios.
  • the Select statement may be in a subquery.
  • the tables in the From list of the Select statement do not have to be real tables or views, they can also be Outer Join clauses.
  • the database also has a Company table with columns CompName, President, City, and State; and the Department table has an added CompName column.
  • Query Q4, derived from query Q3, illustrates adding a Tight Join predicate.
  • each of the two table arguments can be a base table or view or an Outer join.
  • a table set represents the tables in the left argument and another table set represents the tables in the right argument.
  • a table from each set is chosen to supply the column for the fight join predicate and the union of the table sets forms the table set for the encompassing Outer Join.
  • Delete statements and Update statements can also contain predicates created using the tight join process.
  • FIG. 1 For example, suppose an Articles table has columns ArticleName, Reporter. Date, and Text to hold a large number of Newspaper articles The Text column holds the entire text of the article.
  • Query Q5 is an example of a query that can be formed for the example database: Select ArticleName, Date From Articles Where CONTAINS(Text, ‘cherry NEAR pie’) (Q5)
  • Q5 thus returns the names and dates of all articles that contain the word ‘cherry’ somewhere near the word ‘pie’ (perhaps in the same or adjacent sentences).
  • ‘cherry’ and ‘pie’ are character string constants and NEAR is a keyword supported by the database management system.
  • the automated SQL testing tool can generate a query like Q5, but the character string constants would be random length strings of random characters. With rare exceptions these randomly generated constants do not exist as text words in real databases. This causes almost all uses of CONTAINS and other full text predicates to fail, which results in under-exercise of some code paths.
  • FIG. 14 is a flow diagram for generating a sample dictionary of real words from a text column in the database.
  • the dictionary is then used to help generate text strings for statements produced by the process of FIG. 6 .
  • a word dictionary is added as a choice in block 650 , when a character string constant is generated.
  • the process determines the total number of rows, “NumRows,” in the table containing the text column. This parameter can be used to determine how many words (NumWords) should be inserted into a dictionary. NumWords can be determined as an percentage of NumRows, it can be hard coded, or it can be determined in other manners. The invention is not limited to any particular method of determining the number of words to be inserted into the dictionary.
  • the process fetches a random substring from the text column in a randomly selected row.
  • a SQL statement of the form ‘Select SUBSTRING(TextCol, Rand0*Len(TextCol), SubLen) from Table Where KeyCol Rand( )*NumRows.
  • Rand( ) is a random number generator returning values uniformly distributed between 0 and 1.
  • SubLen is the length of the substring retrieved and its value is configurable.
  • 64 is a used as a starting value, however the invention is not limited to any particular value for Sub Len.
  • KeyCol represents a key column of the table.
  • the string returned at block 1404 is scanned from left to right looking for a word.
  • each character is scanned from the left to look for the beginning of a word. Blanks are ignored until a nonblank is found (at char cStart). The scan continues until a blank or punctuation character is found (at char cEnd).
  • the method returns to block 1404 .
  • the process select the characters between cStart and (cEnd-1) as a word and adds it to the Dictionary.
  • a check is made to determine if the dictionary is full (i.e. have NumWords words been inserted). If so, the method terminates, otherwise the method returns to block 1404 .
  • column KeyCol has values 0, 1, . . . NumRows.

Abstract

A test generator produces a set of database query-language statements comprised of randomly chosen elements for testing one or more database management systems on arbitrary databases. The statements are syntactically correct according to the query language, and are semantically correct according to the query language and according to the schema of the target database. A configuration file further specifies parameters of the test statements, in terms of maximum elements, weights of different elements, etc. The generated statements include predicates in which tables in a from clause are tightly joined. In addition, a dictionary of words randomly selected from text columns in a test database is maintained and used to create predicates having words that actually appear in the row data.

Description

    REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation in part of U.S. patent application Ser. No. 09/078,837, filed May 14, 1998; “TEST GENERATOR FOR DATABASE MANAGEMENT SYSTEMS”, which is hereby incorporated by reference herein.
  • FIELD
  • The present invention relates to electronic data processing, and more specifically concerns automated testing of database management systems using statements having a tight join of a plurality of tables.
  • BACKGROUND
  • Relational database management systems (DBMS), such as Microsoft SQL Server, interpret statements written in a database query language such as Structured Query Language (SQL) to create and manage database objects, to insert and update data, and to perform complex, multilevel queries against huge amounts of data. Testing these systems is recognized throughout the industry as a technical challenge of the first magnitude. SQL and similar database-system interpreters are highly complex. For example, they offer sophisticated optimization techniques and execution planning for queries input on the fly; opportunities for arcane design problems are ubiquitous. At the same time, the state space to be tested is gigantic. For a one-gigabyte database, the possible combinations of database configuration and SQL statement to be executed exceeds 102,000,000,000.
  • Libraries of test scripts for relational database systems typically contain thousands or tens of thousands of sample statements which are applied to a test database for comparison of their results with known correct data. Existing libraries are known to be inadequate; most commercial database systems produce a constant and substantial stream of reported bugs. However, the amount of work required to generate larger libraries quickly becomes prohibitive.
  • At the rate of a half hour per hand-written test statement, even a small library consumes more time than does the design of the system that it tests.
  • In the past, developers have employed some stochastic testing at the language level to accelerate database testing. For example, a test-case generator may choose a random mix of hand-generated fixed scripts. Choosing random parameter values in fixed scripts increases the effective number of test cases. These methods still require painstaking human composition and verification of long, multilevel queries. Automated generation of very simple queries considerably speeds up the generation of test cases, but eliminates the more complex test cases where subtle errors lurk.
  • In addition, conventional test systems are effectively limited to a fixed database, or to simple variations on fixed data. In order to construct statements that actually execute properly against the target database, the test system must be internally coded to produce only those statements that match the semantics of the database, the names of the database tables and their columns, the particular data types of each column, and so forth. However, testing on only one set of data obviously restricts the range of the tests that can be performed and thus the errors that will be uncovered. In addition, the use of fixed data for many tests does not permit slanting test runs toward certain kinds of applications, or focusing on the kinds of data or database structures that have been found to produce errors.
  • Also, the length and intricacy of test statements, although desirable for teasing out subtle errors, works against the isolation of bugs which cause those errors. Short, simple statements that produce errors are much more useful for tracking the errors down to particular parts of the DBMS under test.
  • In addition, automatically generating test SQL statements in a random manner has two practical shortcomings. First, when a test SQL statement refers to data in multiple SQL tables, the set of result rows often includes all combinations of data from rows in the tables in the From clause. This is called a Cartesian product of the tables. As a result, the number of output rows is the product of the sizes of the queried tables, which can impose artificial limits in the sizes of tables in the test database.
  • A second shortcoming results from the use of randomly generated character string constants. Some SQL database systems support special searching of large text objects. For example, the text objects might be newspaper articles and the search predicate would specify articles that contained two particular phrases “near” each other. Near could mean anywhere in the same paragraph. Unlike regular character string predicates involving equal, not equal, greater than, etc, the text predicates concentrate on finding words or strings that are in the text. Generating character strings constants with randomly selected characters almost always yields tokens that are not found in the text string. This reduces the effectiveness of random testing since some code paths will not be tested much.
  • The prior art is this field has not satisfied a longstanding need for fast generation and execution of complex test statements for sophisticated database systems that accurately model results sets used in real world applications.
  • SUMMARY
  • The present invention speeds up the generation of database test cases by orders of magnitude. A typical generator running on a personal computer having a single 200 MHz microprocessor outputs 700 SQL statements per second, about a million times faster than a human. The queries are complex and can have multiple nested levels. They have valid semantics as well as valid syntax; that is, they will run correctly on a bug-free database system, using whatever sample database is selected for a test run. The statistical and other features of the test cases are configurable. A test operator may choose the syntactic elements selectable in queries or other statements, the frequency of their use, and parameters such as the maximum subquery depth.
  • Briefly, the invention achieves these and other objectives by reading configuration data containing a set of test parameters, reading the schema of an arbitrary database, then constructing a number of test statements that are syntactically correct for the DBMS being tested, that are semantically compatible with the target database, and that have content and characteristics pursuant to the configuration data. One or more DBMSs under test execute the statements and return result data. Execution errors are detected, as well as result-data differences. Error-producing statements can be converted into greatly simplified statements that provoke the same error, in order to facilitate fault isolation.
  • Generated SQL statements include predicates that are tightly joined in order to avoid results sets that comprise the Cartesian product of the data in the tables. In one aspect of the test system, a From list contains N tables. A list of N sets of table names is then created. Initially, each table set contains one table name from the From list. With each iteration, pairs of table sets are uniformly selected, and a table is uniformly selected from each table set. A column from each selected table is chosen and a predicate equating, or otherwise relating, the two columns is ANDed into the Where clause. The two selected table sets are merged into one, and the two selected table sets are then removed. The process iterates until a single table set remains.
  • A further aspect of the system is that text strings to be included in predicates can be selected from a dictionary. The dictionary is created by sampling the text columns in the target database and extracting a random collection of actual words to place in the dictionary. Words from the dictionary can be randomly interspersed with randomly generated words to form argument values for the full text predicates. The fraction of dictionary words used compared to randomly generated words can be a configuration parameter of the automated SQL testing tool. A separate dictionary is built for each text column in the database. The number of words, to be placed in the dictionary can be a configuration parameter (specified either as a constant or as a percentage of the total number of bytes in the text column).
  • Other features and advantages of the invention, as well as modifications within the scope of the invention, will occur to those having routine skill in the art from the following detailed description, taken in conjunction with the accompanying drawing.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a block diagram of an environment in which the invention can be practiced.
  • FIG. 2 is a block diagram of an illustrative database system to be tested by the invention.
  • FIG. 3 is a flow diagram for a test generator according to the invention.
  • FIG. 4, comprising FIGS. 4A-4F, shows a configuration file used in the process of FIG. 3.
  • FIG. 5 shows an example of a syntax diagram used for constructing a statement in FIG. 3.
  • FIG. 6 is a flow diagram detailing the construction of a test statement according to FIG. 3.
  • FIG. 7 illustrates a state-information file for use in conjunction with an implementation of FIG. 6.
  • FIG. 8 shows an example of a parse tree for a statement constructed by the process of FIG. 6.
  • FIG. 9, comprising FIGS. 9A-9B, illustrates a more typically complex test statement produced by the process of FIG. 6.
  • FIG. 10 is a flow diagram for simplifying a faulty statement according to FIG. 3.
  • FIG. 11 shows a simplified version of the statement of FIG. 8 produced by the process of FIG. 9.
  • FIG. 12 is a flow diagram for providing a tight join for test statements produced by the process of FIG. 6.
  • FIG. 13 provides an exemplary application of the process illustrated in FIG. 12.
  • FIG. 14 is a flow diagram for generating text strings for statements produced by the process of FIG. 6.
  • DETAILED DESCRIPTION
  • FIG. 1 is a high-level block diagram of a conventional client/server computer system 100. Network wiring 110 interconnects a number of personal computers (PCs) 120 to a server 130 via network adapters 121 and 131. Server 130 includes a storage subsystem 132 for holding the large amounts of data in typical enterprise databases. Other system architectures are also suitable environments for the invention; for example, units 120 may be terminals connected to a mainframe or midrange computer 130, or unit 130 may itself comprise a PC coupled to PCs 120 in a peer-to-peer network. For small databases, the entire system 100 may comprise a single PC acting as both client and server. Likewise, file storage may be distributed among a number of different machines. Furthermore, the system can be implemented as a three tier or a multi-tier system. FIG. 1 includes a schematic representations of an external storage medium 133 which may store client and server software for distribution and downloading to clients, and another medium 134, such as a diskette, for offline storage of database tables. Medium 134 can also store instructions and data for the test program of the present invention; the test program can be executed in one or more of the clients 120, or even in server 130.
  • FIG. 2 is a block diagram of a typical conventional client/server database management system 200 capable of operating in system 100, FIG. 2. A client application program 210 executes within each PC 120, under a PC operating system 220 such as a version of the Microsoft Windows operating system, including Windows 95®, Windows 98®, Windows Me®, Windows NT®, or Windows 2000®. Among other functions, client application 210 contains a facility 211 for accepting database queries from a user at a PC 120. In addition to user entries, other application programs 230 executing in some of the PCs 120 may present queries to DBMS client 210, via predefined host-language application-program interfaces (APIs) 231. One of these programs can be the test program of the invention.
  • Within server 130, a DBMS server application 240, such as Microsoft SQL Server, executes under a server operating system 250 such as Microsoft NT. DBMS program 240 provides services for creating, maintaining, and modifying a number of relational databases, exemplified by database 260. Program 240 may employ the file-system services 251 of operating system 250, or may provide its own file system. Operating system 250 could execute a separate instance of the entire DBMS application for each request from a client 210. For greater efficiency, however, program 240 gives each client connection a separate thread 242 in the DBMS kernel. Further, this thread is a native operating-system thread, which carries with it all the Windows NT mechanisms for process memory protection, better access to storage devices, and so forth. Search engine 241 processes queries from individual clients 210 upon tables 261 of a database 260, as described more fully below. It also enforces database integrity with conventional facilities for record locking, atomic transactions, etc. In the Microsoft SQL Server, the interface language between query facility 211 and search engine 241 is Transact-SQL, which provides much of the function of the standard ANSI SQL, 89 and ANSI SQL 92 languages, plus extensions for providing greater flexibility and programmability.
  • The test program may execute query-language statements upon multiple systems, either concurrently or sequentially. In order to differentiate between them, the reference numeral 240′ indicates a second DBMS, which manages the same database 260 as the first system 240 (either at different times or, for some DBMSs, concurrently), or a different database, indicated as 260′.
  • FIG. 3 shows the overall flow of a test generator 300 for producing massive numbers of statements for testing DBMS applications 240 according to the invention. Test program 300 can reside in one of the PC clients 120, FIG. 1, where it may simulate a user, alternatively, it can execute within server 130, or at any other convenient location. Testing a DBMS in this context concerns checking the correctness of its implementation of the language by which it interfaces between databases and users of the databases. These languages are generally referred to as query languages, even though their statements perform in any functions other than querying data in tables; for example, statements can create and modify data tables, insert and modify table data, define transactions for committing data, manage database security, and compile performance statistics. Likewise, the term “query” is usually used as a synecdoche for any statement in an interface language. This embodiment assumes the use of SQL (Structured Query Language), which is actually a family of more or less standardized dialects in common use. Any of these dialects will serve the present purposes. Other query languages such as QBE (Query by Example”) can be substituted easily. Every query language has an explicit syntax specification that defines how to construct valid statements in the language. One aspect of testing the DBMS involves determining whether a statement that is well-formed according to the syntax of the query language does in fact execute without error in the tested DBMS (or, more generally, whether the DBMS reports the proper error status of any statement given to it). Another aspect involves verifying that the results (or lack of results) that a statement produces in that test database are correct. The present embodiment can perform both of these tests.
  • Preliminary blocks 310 begin the entire testing process. Block 311 reads in a configuration file 400 containing a set of parameters for the test procedure. One of the parameters specifies the name of a database 260 to provide the data tables 261. That is, the process is not limited to one or more fixed databases for testing a DBMS, but can employ arbitrary, user-selected target databases. Another parameter is able to specify the name of a different DBMS as a verification DBMS for checking the correctness of returned results. Other parameters of the configuration file will be described below.
  • Block 312 connects test program 300 to test DBMS 240 using the standard Open Database Connection (ODBC) protocol or any other conventional method. Block 313 determines the schema of the test database 240. The schema of a database is the organization of its data. Schemata for relational databases include the names of all data tables, the names, positions, and data types of table columns, restrictions on null or duplicate data values, and so forth. The test-database schema may be derived from any convenient source; it could, for example, be stored in the configuration file. Block 314 symbolizes the syntax specification of DBMS 240. In this implementation, the syntax is built into the code that constructs statements, rather than read in from an external source.
  • Block 320 constructs each statement to be used in the testing process as a parse tree by following the syntax specification of block 313. At each element of the syntax diagram, block 320 inserts a syntactic element which comports with the test-database schema read by block 314. The selection of one of a number of grammatically correct elements is made by a random roll from a seed, guided by probability parameters contained in the configuration file from block 311. The configuration file can, in fact, specify that only certain features, certain syntactical constructs, or certain parts of the database be included in the statements. Actually, block 320 produces statements using a pseudo-random number generator 321, so that the same configuration settings, the same schema, and the same starting seed cause it to produce the same statement deterministically; this allows regression testing of the DBMS. Although one could save the sequence of generated statements for regression testing, the advantage here is that saving only the starting seed can reproduce the entire sequence later. In addition, block 320 follows a set of internal rules that further permit or constrain certain choices at certain points in the parse tree.
  • Block 330 causes test DBMS 240 to execute the statement against database 260. If the configuration file had specified verification testing, then block 330′ causes verification DBMS 240′ to execute the same statement against the database. Each execution may or may not produce a set of data (for a query) or other results from the database. Verify block 331 compares the result sets produced by multiple executions, and produces a verify-error indication if they are not the same. Although the results could be compared in detail, a simpler check usually suffices; for example, block 331 can compare the number of affected rows for a data-modification statement. Block 331 can count the rows in the result sets of a query statement, then generate and compare checksums over the column values in all the rows; this avoids having to sort the result data and compare it one unit at a time. To avoid precision errors, the configuration file can specify a round-off tolerance for numeric fields. Date/time fields are more problematic, because the same value might assume different valid forms; a configuration parameter can specify a common format, or can specify that these fields are not to participate in comparisons.
  • Block 330 also produces an operational error indication when test DBMS 260 fails to execute the statement properly. Besides execution errors (usually including query-language compiler errors) reported by the DBMS, operational errors include lost connections to the database, deadlocks, and system crashes. If a DBMS connection is lost, program 300 attempts to reconnect; if this attempt fails, the program aborts the test run. A system crash requires an automated rapid restart of the server in order to continue the run.
  • Block 332 logs each executed statement, its result data, and any error indications, along with the seed value for each statement. If verification is not being done for the test, logging this data allows the test to be run again at a later time in a pseudo-verification mode with the same DBMS 240, using logged data from the previous run for comparison. This mode is useful for catching suspected intermittent errors.
  • Blocks 340 process data and operational errors resulting from the execution of a statement. If no error occurs, block 341 passes control directly to loop control 350. The extremely long test runs generated by program 300 can result in huge error files of ‘uninteresting’ errors. An uninteresting error is one that is expected to naturally occur in a randomly generated statement. For example, divide-by-zero and overflow errors are uninteresting. Therefore, block 342 filters some of these from the log, or merely counts their occurrences. Additionally, some errors tend to occur in large numbers when they occur at all. For this situation, a configuration-file parameter specifies certain error codes to count or to ignore entirely.
  • Block 343 simplifies statements which have produced errors. Debugging a complex non-procedural program such as a DBMS is greatly facilitated by modifying failed statements to produce simpler versions that still produce the same error. Simplification is particularly advantageous for the long, complex statements produced by program 300. Simplification proceeds by sequentially removing as many elements of the statements as possible while preserving the same error indication. The simplified statement is usually not equivalent to the original statement which means it would not return the same result set had no error in the system occurred. Although one bug might cause a certain error indication in the original statement and a different bug in the simplified one; this has been found to be extremely unlikely. Block 344 records the simplified statement and keys it to the original.
  • Block 350 passes control back to block 320 to generate another statement as long as the test run has not completed. Completion conditions are read from the configuration file. Any convenient measure, such as number of statements, number of errors, and run time, are possible. Block 351 draws up a report of the entire test run. The report can include conventional items such as error listings and statistics concerning the different types of features included in the test statements, or reduced summaries of such items. The report also optionally lists the schema of the test database and the current configuration-file settings. For the pseudo-verify mode described above, it is also useful to include machine readable information on each statement, its result data, and error indications. Program 300 can then run these same statements against the test database on the same test DBMS at a later time, and compare the two runs. Also, a conventional utility program can combine and summarize the results of multiple test runs carried out concurrently on different data processors.
  • This embodiment of program 300 generates and executes an entire suite of test statements in a single run. It is entirely possible, of course, to generate all the test statements in a batch, then to run those statements from a file at a later time. For example, program 300 could execute blocks 310, then loop on block 320 until all statements have been completed. Then a later run could loop on blocks 330-340, and print the report of block 351.
  • FIG. 4 shows a portion of a configuration file 400 containing about a hundred parameters that program 300 employs to control a test run. The format is similar to that of a conventional .INI file for specifying parameters and data to a program. Each line names a parameter and gives it value; comment lines begin with asterisks or dashes. A Program section specifies global parameters controlling the overall operation of a test run, including the database schema of block 313. The parameter fSimplifySQL determines whether or not to auto-simplify faulty statements. A Console section controls data display for real-time tracking of a test run. An SQL file section configures the report file of block 351. A Limits section specifies bounds on database size, overall statement size, subquery depth, number, and so forth. An Allowed section controls the type and content of statements generated by block 320. For example, cSelList:=8 ensures no more than eight output columns in any generated SELECT statement. Many of the parameters occur in pairs naming a maximum and a frequency for statement features or constructs. For example, cGroupByCols=3 and cWTG_GROUPBY=25 specify that no statement has more than three columns in a GROUP BY clause, and that 25% of all statements shall include a GROUP BY clause. A group of parameters near the end of this section specify the mix of different statement types, such as UPDATE, INSERT, DELETE, and SELECT. Parameters such as fUnicode=0 toggle on or off the use of certain data types in generated statements. Frequently, discrepancies arise in block 331 of FIG. 3 because different DBMSs will return the same result data values but in different data types; data-type translation solves such problems without restricting the domain of generated statements. Because transactional integrity is an important aspect of DBMS operation, other parameters control the use of multi-statement transactions and methods, COMMIT and ROLLBACK, to end the transactions. Configuration file 400 not only permits the testing of different systems and databases, but also greatly facilitates the isolation of system bugs. That is, particular features and statement characteristics that produce problems can be emphasized merely by changing a few parameters in the file.
  • Program 300 builds statements in response to parameters from the configuration file, to database structure from the schema, and to the syntax of the query language. Query-language syntax is usually written as a set of productions in a stylized script or diagram format such as Backus normal form. A syntactically valid statement is conceptually constructed by expanding the productions, choosing only one of the allowable paths at every node where the syntax permits alternatives. In this embodiment, individual procedures build major productions such as WHERE and GROUP BY; the smaller productions, such as picking operators in an expression, employ simple case statements.
  • Some of the alternatives can involve repetitions and/or recursive calls. As an illustration of a syntax specification, FIG. 5 shows a partial set of productions 500 for a SELECT statement in a simplified dialect of SQL. The name of each production is printed above its diagram. Rectangular boxes indicate other productions. Rounds and ovals indicate terminal syntactic elements. Small diamonds indicate points at which alternative paths exist.
  • In Select Statement production 510, for instance, block 511 requires that a SELECT statement begin with the word “SELECT”, which can be optionally followed by a qualifier word “ALL” or “DISTINCT” at point 512. Then a Select List Entry 520 can be repeated zero or more times at node 513, separated by commas. Production 520 for a Select List Entry comprises either one or more Column Name syntactic elements 521 separated by Operator elements 522 at choice point 523. Alternatively, point 524 can specify a Function name 525 followed by a parenthesized Column Name 521.
  • The final exit from node 513 following the last Select List expansion requires a “FROM” keyword terminal 514 followed by a Table List Entry 515; the production for this element includes table names, JOIN operators having ON clauses, and other elements, not shown. The other major parts of a SELECT statement are the optional WHERE clause 516, GROUP BY clause 517, HAVING clause 518, and ORDER BY clause 519. Both clauses 516 and 518 include a Search Condition predicate 530. Production 530 illustrates another aspect found in most syntax specifications; a production may include itself as an element, as shown by parenthesized block 530. Among the numerous elements available al: node 531 is an Exists Predicate 540. Production 540 reveals that one of the elements of this predicate is a Select Statement production 510. That is, an entire SELECT statement can be nested inside another SELECT statement; such a nested occurrence is referred to as a subquery, and its containing query or subquery is called an outer query. A column C belonging to a table in an outer query can be referenced in a subquery. The reference to C in the subquery is called a correlated reference.
  • FIG. 6 shows a process 600 for constructing a single test statement according to block 320, FIG. 3, by following syntax diagrams stochastically. One may think of constructing a statement as building a parse tree of the statement; however, routine 600 does not produce any explicit representation of a parse tree. In the following description, file term “roll” means to obtain a number from random number generator 321 and use the number to make a choice among alternatives allowed by certain constraints, with probabilities which might be specified by rules, settings, or other sources; it includes situations where the constraints, etc., might allow only a single choice.
  • First, block 610 rolls to determine the type of statement to be generated. For the SQL query language, the alternatives include SELECT, UPDATE, INSERT, etc. Configuration file 400 lists the desired probabilities of each statement type. Block 611 initializes information described below that is carried along from node to node during the process.
  • Block 620 follows the syntax diagram of the selected statement. That is, at each activation, it chooses the next syntactical element or set of alternative elements in the diagram as the current node. In diagram 500, for example, block 620 would first choose element 511, then the alternatives presented at point 512, then, depending upon a roll outcome, neither or one of the elements 521 and 525, etc., then onward through diagram 520. Returning to diagram 510, a roll at point 513 sends block 620 through element 520 again, or on to element 514. Some elements might be visited recursively and/or repetitively as block 620 travels upward and downward through the diagrams 500. Block 621 outputs a representation of the statement to program 300 as the statement is generated. If the statement will be executed, block 621 also places a copy in a buffer.
  • In general, process 600 works on one syntactic element at a time to build one node of the parse tree. At certain places, however, it is necessary to gather preliminary information and to make certain preliminary choices involving elements other than those at the current node in the diagram. For example, the names of the tables to be used in a SELECT query need not be determined syntactically until block 620 reaches element 515 in FIG. 5. However, the choice of columns at previously encountered element 520 could prow3ke inconsistencies with certain constraints, such as parameter values in configuration file 400 or the schema of the database. Therefore, block 630 obtains any preliminary information specified for the current node, block 631 rolls any choices specified by the information, and block 632 records choices that might be useful or might affect choices further along the tree. Continuing the above example, when element 520 becomes the current node in a SELECT statement, block 630 retrieves the list of tables from the schema and the maximum number of tables in a join parameter from configuration file 400. Block 631 chooses a list of particular tables randomly within the parameter limit, and reads the names and data types of all columns belonging to these tables. Block 632 records the names of these columns, because this choice affects the alternatives available to the subsequent element 515, and also to other elements, such as the Column List of GROUP BY clause 517. As an added optimization, block 632 also records the data types of these columns in order to avoid having to look them up again while processing subsequent elements involving functions of those columns.
  • Block 640 gets information pertaining to the current node that can affect the alternatives available at that node. State information such as the identity of the present element in the syntax diagram obviously affects this choice. Certain parameter values in configuration file 400 might be relevant to the current node; for example, the maximum subquery depth might bar the choice of element 540 at point 531 in FIG. 5, because this element always generates a subquery at element block 510. Some of the preliminary choices might affect the current choices; this includes new information recorded by block 632 for the current node and for any higher node that may affect the current choice. For instance, the preliminary choice of less than all tables, made at element 515, precludes later nodes from choosing columns in other tables of the database schema.
  • Block 641 presents a cumulative history of previous choices made for any previous nodes in the parse tree. In building a tree by doing a walk of the tree, previous Choices can affect subsequent alternatives at the current node. For example, the data type of an expression is chosen before the expression is generated so the data type information is passed down the tree. This is true even when the alternatives might be syntactically correct; for instance, an ORDER BY entry can be an integer to index the select list but the integer value can not exceed the number of elements in the Select List. The difference between blocks 640 and 641 is that the first restricts alternatives at the current node based upon determinations made for nodes higher (toward the root) in the constructed parse tree, while block 641 carries choices made at lower (toward the leaves) nodes upward to restrict choices when the current node is higher in the tree and choices made at sibling nodes in other subtrees of the parse tree. Stated another way, block 641 concerns all the node choices made up to the current node that are not in the path from the current node to the root. Carrying the results of choices to higher nodes is one of the most important factors in constructing semantically correct test statements for arbitrary databases according to the invention. Without a facility for carrying previous choices upward in the tree, process 600 would have to cleave closely to a predefined fixed database in order to construct statements consistent with the semantics of the database.
  • Block 642 makes available a set of rules written for particular nodes. These rules are of two types. Mandatory rules require the performance of certain actions. For example, some points in a syntax diagram require that a column reference be generated; thus any potential alternatives at the current node that do not result in a column reference must be excluded. Optional rules permit certain actions. Such a rule might allow the generation of a column reference to a correlated column in a subquery at the current node. Rules can be hard-coded into a program that executes process 600, stored in an external file, or implemented in any other convenient manner.
  • Block 650 assembles a list of alternatives available at the current node. These alternatives arise from the set of elements specified at the current node of the syntax diagram, and are potentially modified or restricted by some or all of the information obtained in blocks 640-642. Block 651 rolls the assembled alternatives to produce a random choice of a syntactic element for the current node. (As a shorthand notation, the term ‘roll’ means to choose among a set of alternatives in accordance with a random or pseudorandom number; the probabilities of each alternative can be equal or weighted.) This can include a terminal value; e.g., a roll might produce the numeric value 10000 where the current node requires an item of numeric data. Some of the alternatives might have probabilities or ranges specified by weighting parameters in the configuration file, as mentioned. Although the roll outcome could be derived from a non-predictable random source, such as hashing the time of day, using a pseudo-random roll allows a seed value recorded in the configuration file to reproduce the action of block 651 deterministically if desired for regression testing or other purposes. For degenerate nodes where only one alternative exists, block 651, or blocks 640-650, can be entirely bypassed to save time.
  • Block 652 records the choices made in the history obtained by block 641. Block 653 updates the information obtained by block 640, based upon the current-node choices. Block 654 updates the statement according to the choices made in block 651. Normally, a choice at each current node adds a term to the statement's parse tree. However, it is sometimes advantageous to delay this action. For example, column references are constructed for a GROUP BY clause 517, FIG. 5, but later choices can generate expressions and intermix them randomly with the column references before outputting the clause as part of a SELECT statement or subquery.
  • In practice, process 600 can be implemented as a number of individual functions called recursively on a parse tree. As an example, a C function GenSelect ( ) produces a single SQL statement 510 at any level of a parse tree. Briefly, this function first rolls to determine the list of tables in the FROM clause, within the configured limit. If this query must return a particular data type, it checks to ensure that at least one listed table contains a column of that type. Next, the function builds a list of data types for all columns in the selected tables. Another roll determines how many elements the select element 520 will contain. The next roll determines whether the query will include a DISTINCT clause, and/or will use the DISTINCT qualifier in aggregate functions. A roll includes or omits a GROUP BY clause, and, in the former case, rolls to build its column references—but does not yet output the clause.
  • The next step is to output the “SELECT” keyword to the tree, and start following syntax diagram 500. A sequence of rolls generates expressions for syntactic element 520 which are then output. The keyword “FROM” and the list of tables in the FROM clause is then output to the parse tree. If outer joins are configured, a sequence of rolls generates their ON-clause predicates. A roll includes or omits a WHERE clause 516, and performs further rolls to build it. If a predicate involves a subquery, GenSelect ( ) produces certain rules (e.g., return no more than one row of data), and calls itself recursively to build the subquery. The function then produces expressions for the GROUP BY list, and outputs the entire GROUP BY clause. A further roll possibly includes and builds a HAVING clause, building any subqueries and their rules as above. Finally, a roll possibly includes and builds an ORDER BY clause, then rolls to intermix expressions and index references to entries in the Select List.
  • Another function, GenExpr ( ), builds expressions for GenSelect ( ) and for other functions. It outputs a constant or a column reference, or calls itself recursively to generate operators or functions that take expressions as arguments. While the design of such a function is conventional, this particular expression builder illustrates the use of rules according to the invention. A call to GenExpr( ) includes two parameters. TableInfo holds the names and data types of table columns, as described above. ExprInfo contains the explicit rules that the expression must follow, and includes state information that influences the function's choices for expression terms. FIG. 7 lists the parameters of the ExprInfo rules and their meanings. Element names beginning with ‘f’ are binary flags. If set, the corresponding rule must be followed The rule element DataType, e.g., is a mandatory rule enforcing a particular data type resulting from the evaluation of the expression, while fLocalFunctions is an optional rule allowing aggregate functions such as AVG( ) for any column contained in a table in the current FROM clause.
  • FIG. 8 shows an example of a parse tree 800 for a very short SELECT statement, as generated by process 600. The root of the tree is the SQL keyword “SELECT”, 810. The next level of the tree has branches to three nodes. Although block 820 of the select list 520, FIG. 5, is grammatically the first node, process 600 proleptically produces a pool of possible table and column names, as described above. Block 820 specifies that the name column of the database's employee table is the first item in the list. The second item in the Select List is an expression, the sum 821 of columns named salary 822 and commission 823. Nodes 822 and 823 are leaf nodes; that is, their contents specify terminal elements in the syntax of a SELECT statement. The second high-level tree branch leads to node 830, the keyword “FROM” for table list 515. The only table included in this statement is employee, at leaf node 831. Again, process 600 had chosen this table, before actually outputting nodes 820-823; therefore, the column nodes in 820-823 were restricted to be columns of the employee table.
  • The third high-level branch specifies a WHERE clause at node 840. Nodes 841-847 constitute a predicate generated by a call to a function GenPred( ). Node 841 defines this predicate as a logical AND of two subpredicates. Nodes 842-844, generated by a recursive call to GenPred( ), define the first subexpression as the condition that the value of the salary column exceed the numeric constant 10000. Nodes 843 and 844 are generated by calls to GenExpr( ). The second sub predicate, nodes 845-847, specifies rows of the employee table where the department column contains the string constant “sales”. Had the state information produced during the construction of node 831 not noted that at least one of the columns of the chosen employee table has a numeric data type, all arithmetic operators would have been dropped from the list of alternatives during the construction of the expression at nodes 821-823, which would have precluded the choice of addition operator 821. (Relational operator 842 and equality operator 845 are compatible with most data types.)
  • Process 600 builds statements in the form of parse trees such as 800. However, program 300 normally outputs test statements in the same form that a user or another program would input them to DBMS 240, FIG. 2. Therefore, output block 621 in FIG. 6 outputs each statement as an equivalent character-based representation. Parse tree 800, for example, becomes the statement:
      • SELECT name, salary+commission
        • FROM employee
        • WHERE (salary>10000) AND (department=‘sales’)
  • FIG. 9 shows a more typically complex SQL statement generated by process 600 for a publishing-company database. Statement 900 nests subqueries up to five deep, and the inner queries reference correlated columns in the outer queries. Although some of the requests may seem a bit bizarre (e.g., royalty amounts expressed in radians), the statement syntactically follows the DBMS query language, and the requested data comports semantically with the rules of SQL and with the schema of the target database. Indeed, some of the more prolix constructs, such as adjacent minus signs and redundant parentheses, frequently provoke design errors that more usual statements overlook. One might think that such complex statements would rarely produce any result data at all. Experience has shown, however, that about 50% of non-error SELECT statements generated by process 600 do return at least one row of data. The maximum join size together with the database size and structure (schema) strongly influence the number of returned rows. For large databases, the maximum join size and the maximum subquery depth should be kept low in order to avoid excessive run times and system limits. In a sample test run, ten different clients 120, FIG. 1, concurrently executed 25,000 SQL statements equally distributed over SELECT, INSERT, UPDATE, and DELETE types. Expected errors (mostly deadlocks) occurred in 3,464 statements. Unexpected errors (bugs) having two different error codes occurred in a total of eighteen statements.
  • FIG. 10 shows a process 1000 for simplifying an error statement (that is, one which has produced an error indicating a fault in DBMS 240) according to block 342, FIG. 3. Because SQL and other query languages are non-procedural, and databases frequently contain gigabytes of data, the query compiler of DBMS 240 attempts to reduce processing time and memory usage with complex access plans involving many steps. Debugging the complex plans step by step is extremely tedious and difficult. Therefore, finding a simple statement having the same error as a much more complex statement aids greatly in isolating the cause of the error.
  • Block 1010 first reads in the error statement detected by block 341 in the format of a parse tree as shown in FIG. 8. Block 1011 records an indication denoting the particular error that occurred; usually, this is a code produced by search engine 241, although it can also be a signal indicating a connection loss, or some other indication. Block 1012 initializes process 1000 to the root node of the parse tree (e.g., node 810, FIG. 8).
  • Block 1020 walks the parse tree of the error statement, using any conventional method. Block 1020 chooses certain nodes whose subtree is a candidate for removal. Block 1030 removes the entire subtree of the current node. (For some cases, such as expressions involving two operands such as ‘A+B’, block 1030 first removes subtree A and also the ‘+’. Then it removes subtree B and also the ‘+’.) Block 1031 then reexecutes the truncated statement on the search engine. If the search engine returns the same error indication, block 1032 passes control to block 1020 to determine whether this truncated statement can be further simplified by removing additional subtrees. If the reexecution results in no error code, or in a different error code, block 1033 restores the subtree removed by block 1030 before passing control back to block 1020. When block 1020 has visited all branches of the parse tree, block 1021 translates the truncated tree into the form of a simplified statement and outputs it to block 344, FIG. 3.
  • Using the example in FIG. 8, block 1020 first visits the branch to node 820, to remove the subtree including nodes 820-823, then the branch to node 821, to remove nodes 822-823, then to remove 822 alone, then to 823 alone. Next, the branch to node 830 is removed, and that subtree is followed to its leaf 831. Finally, the branch to node 840 is removed, and its subtree is visited. Although the parse tree of the original statement could be processed in the opposite direction, walking the branches from the root node toward the leaf nodes removes the highest possible subtrees first, thus isolating the miscreant subtree more quickly. Also, block 1020 need not attempt to remove each possible subtree of the parse tree. For example, removing the entire subtree 820-823 would result in a simplified statement having incorrect syntax, because the entire select list would be missing from the SELECT statement. Block 1032 would then reject the resulting syntax-error code, causing block 1033 to restore the subtree. Removing the subtree containing nodes 821-823, however, leaves the statement in a syntactically correct form. (FIG. 8 has been simplified for this short example. For more complex statements, removing a subtree does not remove a suffix of the Select List. Select List entries are not removed, but rather only elements of their expressions. In expressions with two operands, the operator is removed along with the subtree of one of the operands.)
  • A more intelligent block 1020 understands at least some of the formalities of the target query language, and passes over branches which are known to be syntactically required or to possess other properties rendering them unsuitable for deletion. A number of heuristic optimizations have been incorporated into the statement simplifier. For example, it performs trial deletes from the top of the tree down; this alone eliminates the processing of most of the subtrees. The simplifier does not try all possible deletions. For example, it does not lop items from the Select List or from the ORDER BY list. It does not always plumb the deepest level. It never back-tracks. Once it tries a delete and finds it doesn't work, it never tries the same delete again, although there are cases where other successful deletes enable a previously tried one.
  • FIG. 11 shows a simplified statement generated from the statement of FIG. 8 by process 1000. Such drastic parings of the parse tree are not uncommon. Modifying elements of the simplified statement for further fault isolation is now feasible, whereas modifying portions of the full statement of FIG. 4 would be a daunting task.
  • The above-described systems and methods provide a fast and efficient mechanism to generate SQL statements that can be used to test database software. In some embodiments of the invention, the SQL statements generated using the systems and processes described above are created using a tight join. As is known in the art, an SQL join is the combining of two (or more) SQL tables, often with a predicate that relates rows in different SQL tables. For example, suppose SQL table Employee has columns Name, DeptNumber, and Salary and contains 1000 rows. Also suppose that SQL table Department has columns DeptNumber, DeptName, Manager, and Location and contains 100 rows. The SQL query Q1:
    Select Employee.Name, Employee.Salary, Department.Location From Employee, Department  (Q1)
    is an example of a join of the Employee and Department tables. The result of executing the above query is a table with 3 columns and 100,000 rows representing all combinations (1000×100) of rows in both tables. The result table is the size of the Cartesian product table. In joins of tables where each table has many rows, the Cartesian product can grow quite large, and it can take a large amount of time to return the complete result set.
  • If a predicate is added to the query to limit the result set to those rows where there is a match on Employee and Department rows corresponding to the same department, then the result table will have at most 1000 rows in it with each row corresponding to one employee. The query Q2 below illustrates such a query:
    Select Employee.Name, Employee.Salary, Department.Location From Employee, Department
    Where Employee.DeptNumber=Department.DeptNumber  (Q2)
    The query in Q2 filters the results to exclude those employees with a DeptNumber value that does not match a value in Department. The query in Q2 is thus an example of a tight join: the two tables in the From list of the query are joined with a tight join predicate of the form Table1.Column1=Table2.Column2. Further “ANDing” of predicates results in an even more restrictive join with correspondingly fewer rows returned in the results set. The query in Q3 is an example of adding further predicates to further filter the results set:
    Select Employee.Name, Employee.Salary, Department.Location From Employee, Department
    Where (Employee. Salary>10000 or Employee. Name< >Department. Location) AND Employee. DeptNumber=Department. DeptNumber  (Q3)
  • FIG. 12 is a flow diagram of a process 1200 for providing a tight join for test statements produced by the systems and processes described above in reference to FIGS. 1-11. The process begins at block 1202, where a SQL Select statement is generated as described above. The Select statement will have a list of N tables in the From list. Note that if N=1, the process exits, as there can be no join formed based on a single table.
  • At block 1204, a plurality of table sets is created, where each table set initially contains one table in the From list. Thus initially, the I-th table set will contain the I-th table in the From list.
  • Next, at block 1206, a pair of table set TSj and TSk are selected from the plurality of table sets. In some embodiments, the table set pairs are selected uniformly, that is, randomly. As those of skill in the art will appreciate, other selection algorithms can be used. For example, the least recently used table set could be selected. The invention is not limited to any particular mechanism for selecting a table set pair.
  • Then, at block 1208, a table T1 and Tm are selected from table sets TSj and TSk respectively. As the process iterates, the table sets will contain more and more tables. In some embodiments of the invention, the selection of a particular table from a table set is uniform. In alternative embodiments, the tables are selected using other algorithms known to those of skill in the art.
  • At block 1210, columns Cx and Cy are selected from tables T1 and Tm respectively. In one embodiment of the invention, the columns are selected uniformly. In an alternative embodiment, the columns are selected according to whether they have a compatible data type such that they can be logically compared with little or no type conversion (i.e. conversion of a numeric to a character string or vice versa). In a further alternative embodiment, a list of column pairs that have the same or similar column names is created. A column pair can be considered similar if the pair has a common prefix or suffix, are phonetically similar, or exhibit a common pattern. A column pair is then chosen uniformly from the list of similar column pairs. Selection of similar column names is desirable, because it is often the case that similar column names represent foreign keys relating on table to another, and thus emulates the manner in which select statements will be used by actual customers.
  • After the columns have been selected, at block 1212 a predicate is formed using the two column names. The main implementation is for the predicate to perform a logical comparison to determine if the data value for column Cx is equal to that for column Cy, but other types of comparisons are possible that preclude Cartesian products. The predicate is “ANDed” into the existing Where clause for the SQL statement.
  • At block 1214, the tables in table sets TSj and TSk are merged into a single table set, and TSj and TSk are deleted. A decision block 1216 determines if more than one table set remains after the merging. If so, the process proceeds to block 1206 to repeat the predicate creation process. If not, the Where clause is complete and the process ends. The Where clause that is generated represents a tight join of the tables in the From clause, and restricts the results set such that a Cartesian product of the rows in the tables is avoided.
  • FIG. 13 provides an exemplary scenario of the application of the process described in FIG. 12, in which a SQL Select statement is transformed to include a Tight join. In the example, the original query 1302.0 is shown in along with an initial five table sets 1304.0 each containing one table from the From list. In the first iteration of process 1200 above, table sets {T3} and {T5} are used to form the predicate T3.c=T5.c, which is added to form SQL statement 1302.1. Table sets {T3} and {T5} are merged into a single table set {T3,T5} resulting in table set 1304.1. Each iteration of process 1200 reduces the number of table sets 1304.2-1304.4 by one until only one table set 1304.4 remains. The SQL statements that are created in each iteration are represented in 1302.2-1302.4.
  • It should be noted that the examples presented above illustrate a top-level Select statement with a From clause consisting of simple tables. However, the invention is not so limited. The Tight Join process can also applied in exactly the same way to other scenarios. For example, the Select statement may be in a subquery. Alternatively, the tables in the From list of the Select statement do not have to be real tables or views, they can also be Outer Join clauses. For example assume that the database also has a Company table with columns CompName, President, City, and State; and the Department table has an added CompName column. Query Q4, derived from query Q3, illustrates adding a Tight Join predicate.
    Select Employee.Name, Employee.Salary, Department.Location, Company.State From Employee, Department Outer Left Join Company ON Department.CompName=Company. CompName
    Where (Employee.Salary>10000 or Employee.Name< >Department.Location) AND Employee.DeptNumber=Department.DeptNumber
  • Note that the Outer Joins can nest, each of the two table arguments can be a base table or view or an Outer join. In all cases, a table set represents the tables in the left argument and another table set represents the tables in the right argument. A table from each set is chosen to supply the column for the fight join predicate and the union of the table sets forms the table set for the encompassing Outer Join.
  • In addition, the tight join process described above can be applied to statements other than the Select statement. For example, Delete statements and Update statements can also contain predicates created using the tight join process.
  • Further embodiments of the invention include processes that create predicates involving test strings. Many database management systems, including Microsoft SQL Server have a ‘Full Text’ feature where large text strings in the database can be searched with special predicates pertaining to text search. For example, suppose an Articles table has columns ArticleName, Reporter. Date, and Text to hold a large number of Newspaper articles The Text column holds the entire text of the article. Query Q5 is an example of a query that can be formed for the example database:
    Select ArticleName, Date
    From Articles
    Where CONTAINS(Text, ‘cherry NEAR pie’)  (Q5)
  • Q5 thus returns the names and dates of all articles that contain the word ‘cherry’ somewhere near the word ‘pie’ (perhaps in the same or adjacent sentences). In Q5, ‘cherry’ and ‘pie’ are character string constants and NEAR is a keyword supported by the database management system. The automated SQL testing tool can generate a query like Q5, but the character string constants would be random length strings of random characters. With rare exceptions these randomly generated constants do not exist as text words in real databases. This causes almost all uses of CONTAINS and other full text predicates to fail, which results in under-exercise of some code paths.
  • FIG. 14 is a flow diagram for generating a sample dictionary of real words from a text column in the database. The dictionary is then used to help generate text strings for statements produced by the process of FIG. 6. A word dictionary is added as a choice in block 650, when a character string constant is generated. First, at block 1402, the process determines the total number of rows, “NumRows,” in the table containing the text column. This parameter can be used to determine how many words (NumWords) should be inserted into a dictionary. NumWords can be determined as an percentage of NumRows, it can be hard coded, or it can be determined in other manners. The invention is not limited to any particular method of determining the number of words to be inserted into the dictionary.
  • Next, at block 1404 the process fetches a random substring from the text column in a randomly selected row. In one embodiment of the invention, a SQL statement of the form ‘Select SUBSTRING(TextCol, Rand0*Len(TextCol), SubLen) from Table Where KeyCol=Rand( )*NumRows. Rand( ) is a random number generator returning values uniformly distributed between 0 and 1. SubLen is the length of the substring retrieved and its value is configurable. In one embodiment, 64 is a used as a starting value, however the invention is not limited to any particular value for Sub Len. KeyCol represents a key column of the table.
  • Then at block 1406 the string returned at block 1404 is scanned from left to right looking for a word. In one embodiment, each character is scanned from the left to look for the beginning of a word. Blanks are ignored until a nonblank is found (at char cStart). The scan continues until a blank or punctuation character is found (at char cEnd). At block 1407, if no word is found the method returns to block 1404.
  • Next, at block 1408, the process select the characters between cStart and (cEnd-1) as a word and adds it to the Dictionary. At block 1410 a check is made to determine if the dictionary is full (i.e. have NumWords words been inserted). If so, the method terminates, otherwise the method returns to block 1404.
  • The process makes random probes to obtain words for the dictionary for use in predicates involving text strings. In one embodiment of the invention, column KeyCol has values 0, 1, . . . NumRows. In alternative embodiments, the predicate can be “KeyCol>=MinKey+Rand0*(MaxKey−MinKey) where MinKey and MaxKey denote the minimum and maximum values of the key, respectively.
  • Although specific embodiments have been illustrated and described herein, it will be appreciated by those of ordinary skill in the art that any arrangement which is calculated to achieve the same purpose may be substituted for the specific embodiments shown. This application is intended to cover any adaptations or variations of the present invention.
  • For example, while the embodiments of the invention have been described as executing within a test environment for a relational database management system. The systems and methods of the invention could be applied to object oriented database using Object Query Langauges (OQL) as well.
  • The terminology used in this application is meant to include all of these environments. Therefore, it is manifestly intended that this invention be limited only by the following claims and equivalents thereof.

Claims (9)

1-26. (canceled)
27. A computerized method for generating a dictionary of words for use in a SQL statement generator, the method comprising:
means for determining the number of rows in a table having a text column;
means for selecting a row from the table;
means for selecting a substring comprising a word from the text column of the selected row; and
means for inserting the word into a dictionary associated to the table.
28. The computerized method of claim 27, wherein a size of the dictionary is a function of the number of rows in the table.
29. The computerized method of claim 27, wherein the row is selected randomly.
30. The computerized method of claim 27, wherein the substring is selected randomly.
31. A system for generating a dictionary of words for use in a SQL statement generator, comprising:
a table having a number of rows with a text column, wherein a row is selected from the table;
a substring that comprises a word from the text column, wherein the substring is selected from the selected row; and
a dictionary associated to the table, wherein the word is inserted into the dictionary.
32. The system of claim 31, wherein a size of the dictionary is a function of the number of rows in the table.
33. The system of claim 31, wherein the row is selected randomly.
34. The system of claim 31, wherein the substring is selected randomly.
US10/978,965 1998-05-14 2004-11-01 Test generator for database management systems providing tight joins Abandoned US20050065948A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/978,965 US20050065948A1 (en) 1998-05-14 2004-11-01 Test generator for database management systems providing tight joins

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US09/078,837 US6138112A (en) 1998-05-14 1998-05-14 Test generator for database management systems
US09/677,684 US6581052B1 (en) 1998-05-14 2000-10-02 Test generator for database management systems
US10/411,911 US6826558B2 (en) 1998-05-14 2003-04-10 Test generator for database management systems providing tight joins
US10/978,965 US20050065948A1 (en) 1998-05-14 2004-11-01 Test generator for database management systems providing tight joins

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10/411,911 Continuation US6826558B2 (en) 1998-05-14 2003-04-10 Test generator for database management systems providing tight joins

Publications (1)

Publication Number Publication Date
US20050065948A1 true US20050065948A1 (en) 2005-03-24

Family

ID=28677786

Family Applications (4)

Application Number Title Priority Date Filing Date
US09/677,684 Expired - Lifetime US6581052B1 (en) 1998-05-14 2000-10-02 Test generator for database management systems
US10/411,911 Expired - Lifetime US6826558B2 (en) 1998-05-14 2003-04-10 Test generator for database management systems providing tight joins
US10/978,965 Abandoned US20050065948A1 (en) 1998-05-14 2004-11-01 Test generator for database management systems providing tight joins
US10/980,734 Expired - Lifetime US7007007B2 (en) 1998-05-14 2004-11-03 Test generator for database management systems providing tight joins

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US09/677,684 Expired - Lifetime US6581052B1 (en) 1998-05-14 2000-10-02 Test generator for database management systems
US10/411,911 Expired - Lifetime US6826558B2 (en) 1998-05-14 2003-04-10 Test generator for database management systems providing tight joins

Family Applications After (1)

Application Number Title Priority Date Filing Date
US10/980,734 Expired - Lifetime US7007007B2 (en) 1998-05-14 2004-11-03 Test generator for database management systems providing tight joins

Country Status (1)

Country Link
US (4) US6581052B1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140101097A1 (en) * 2012-10-09 2014-04-10 Sap Ag Template based database analyzer
US9015165B1 (en) * 2014-03-03 2015-04-21 Michael L. Hamm Text-SQL relational database
US9152630B1 (en) * 2011-10-26 2015-10-06 Intuit Inc. Modified database transaction scopes during software testing

Families Citing this family (86)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6581052B1 (en) * 1998-05-14 2003-06-17 Microsoft Corporation Test generator for database management systems
JP2002007169A (en) * 2000-06-23 2002-01-11 Nec Corp System for measuring grammar comprehension rate
US20020107653A1 (en) * 2001-02-02 2002-08-08 Kraffert Mark J. Sharing data files in a test environment
US7904454B2 (en) * 2001-07-16 2011-03-08 International Business Machines Corporation Database access security
CA2355418A1 (en) * 2001-08-16 2003-02-16 Ibm Canada Limited-Ibm Canada Limitee A schema for sql statements
US7069263B1 (en) * 2002-02-19 2006-06-27 Oracle International Corporation Automatic trend analysis data capture
US6880149B2 (en) * 2002-04-01 2005-04-12 Pace Anti-Piracy Method for runtime code integrity validation using code block checksums
US6983275B2 (en) * 2002-04-16 2006-01-03 International Business Machines Corporation Optimizing database query by generating, determining the type of derived predicate based on monotonicity of the column generating expression for each remaining inequality predicate in the list of unexamined predicates
US8301657B1 (en) * 2002-05-09 2012-10-30 Teradata Us, Inc. Set-level database access for performing row-sequential operations
US7519948B1 (en) * 2002-12-26 2009-04-14 Judson Ames Cornish Platform for processing semi-structured self-describing data
US20040193575A1 (en) * 2003-03-25 2004-09-30 Chia-Hsun Chen Path expressions and SQL select statement in object oriented language
US7143107B1 (en) 2003-06-26 2006-11-28 Microsoft Corporation Reporting engine for data warehouse
US8856163B2 (en) * 2003-07-28 2014-10-07 Google Inc. System and method for providing a user interface with search query broadening
CN1293480C (en) * 2003-08-11 2007-01-03 摩托罗拉公司 Storage unit for text input predication and user dictionary
US7337176B1 (en) * 2003-08-29 2008-02-26 Sprint Communications Company L.P. Data loading tool for loading a database
US7665067B2 (en) * 2003-09-15 2010-02-16 Cadence Design (Israel) Ii Ltd. Method and system for automatically creating tests
US7426522B2 (en) * 2003-09-23 2008-09-16 International Business Machines Corporation Object oriented query path expression to relational outer join translator method, system, article of manufacture, and computer program product
US7188098B2 (en) * 2003-09-24 2007-03-06 International Business Machines Corporation Query transformation for union all view join queries using join predicates for pruning and distribution
JP2005122560A (en) * 2003-10-17 2005-05-12 Fujitsu Ltd Program for detecting deadlock beforehand
US7437362B1 (en) * 2003-11-26 2008-10-14 Guardium, Inc. System and methods for nonintrusive database security
US7636710B2 (en) * 2004-03-04 2009-12-22 Symantec Operating Corporation System and method for efficient file content searching within a file system
US7246116B2 (en) * 2004-04-22 2007-07-17 International Business Machines Corporation Method, system and article of manufacturing for converting data values quantified using a first measurement unit into equivalent data values when quantified using a second measurement unit in order to receive query results including data values measured using at least one of the first and second measurement units
US7487396B2 (en) * 2004-10-15 2009-02-03 Broadcom Corporation System and method to locate and correct software errors within a protocol stack for wireless devices
US7580923B2 (en) * 2005-05-10 2009-08-25 Microsoft Corporation Binding for multi-part identifiers
US7761430B2 (en) * 2005-05-12 2010-07-20 Microsoft Corporation Verification of cross domain data system query results
US7970788B2 (en) * 2005-08-02 2011-06-28 International Business Machines Corporation Selective local database access restriction
US7933923B2 (en) * 2005-11-04 2011-04-26 International Business Machines Corporation Tracking and reconciling database commands
US7882121B2 (en) * 2006-01-27 2011-02-01 Microsoft Corporation Generating queries using cardinality constraints
US8225285B2 (en) * 2006-04-03 2012-07-17 International Business Machines Corporation Generating a test suite to test support for routines
US8656374B2 (en) * 2006-06-16 2014-02-18 Business Objects Software Ltd. Processing cobol data record schemas having disparate formats
US7640261B2 (en) * 2006-06-16 2009-12-29 Business Objects Software Ltd. Apparatus and method for processing data corresponding to multiple COBOL data record schemas
US7702616B1 (en) * 2006-06-21 2010-04-20 Actuate Corporation Methods and apparatus for processing a query joining tables stored at different data sources
US8539474B2 (en) * 2006-09-28 2013-09-17 International Business Machines Corporation Method and system for management of interim software fixes
US7680782B2 (en) * 2006-10-18 2010-03-16 International Business Machines Corporation Method to generate semantically valid queries in the XQuery language
GB2443264A (en) * 2006-10-27 2008-04-30 Ntnu Technology Transfer As Integrity checking method for a device in a computer network, which controls access to data; e.g. to prevent cheating in online game
US8141100B2 (en) 2006-12-20 2012-03-20 International Business Machines Corporation Identifying attribute propagation for multi-tier processing
US8495367B2 (en) 2007-02-22 2013-07-23 International Business Machines Corporation Nondestructive interception of secure data in transit
US7953674B2 (en) * 2007-05-17 2011-05-31 Microsoft Corporation Fuzzing system and method for exhaustive security fuzzing within an SQL server
US8726177B2 (en) * 2007-07-18 2014-05-13 Sas Institute Inc. Systems and methods for generating a database query using a graphical user interface
US8473915B2 (en) * 2007-08-03 2013-06-25 International Business Machines Corporation Coverage analysis tool for testing database-aware software applications
US8276117B2 (en) * 2007-08-03 2012-09-25 International Business Machines Corporation Displaying and refactoring programs that include database statements
JP5048417B2 (en) * 2007-08-07 2012-10-17 株式会社富士通ビー・エス・シー Database management program and database management apparatus
US8903801B2 (en) * 2007-09-14 2014-12-02 Oracle International Corporation Fully automated SQL tuning
US8341178B2 (en) * 2007-09-18 2012-12-25 Oracle International Corporation SQL performance analyzer
US8700608B2 (en) * 2007-10-17 2014-04-15 Oracle International Corporation SQL execution plan verification
US8838651B2 (en) * 2008-01-10 2014-09-16 International Business Machines Corporation Database system testing
US8261326B2 (en) 2008-04-25 2012-09-04 International Business Machines Corporation Network intrusion blocking security overlay
US7937385B2 (en) * 2008-05-05 2011-05-03 International Business Machines Corporation Obtaining a plan for executing a query in a relational database
US9720971B2 (en) * 2008-06-30 2017-08-01 International Business Machines Corporation Discovering transformations applied to a source table to generate a target table
WO2010014917A1 (en) * 2008-07-31 2010-02-04 Telcordia Technologies, Inc. Versioning relational database disjoint records
US8266029B2 (en) * 2009-09-04 2012-09-11 Hartford Fire Insurance Company System and method for managing data relating to investments from a variety of sources
US8140476B2 (en) * 2009-12-16 2012-03-20 International Business Machines Corporation Statistical quality monitoring and enhancement
US9104484B2 (en) * 2010-04-21 2015-08-11 Salesforce.Com, Inc. Methods and systems for evaluating bytecode in an on-demand service environment including translation of apex to bytecode
US8843893B2 (en) * 2010-04-29 2014-09-23 Sap Ag Unified framework for configuration validation
US20120041989A1 (en) * 2010-08-16 2012-02-16 Tata Consultancy Services Limited Generating assessment data
US8392399B2 (en) * 2010-09-16 2013-03-05 Microsoft Corporation Query processing algorithm for vertically partitioned federated database systems
US8438153B2 (en) * 2010-10-11 2013-05-07 Hewlett-Packard Development Company, L.P. Performing database joins
AU2012203333A1 (en) 2011-06-15 2013-01-10 Agile Software Pty Limited Method and apparatus for testing data warehouses
US9268828B2 (en) 2011-06-15 2016-02-23 Sas Institute Inc. Computer-implemented systems and methods for extract, transform, and load user interface processing
US9104720B2 (en) * 2012-06-28 2015-08-11 International Business Machines Corporation Generation of technical description of report from functional description of report
US20140006459A1 (en) * 2012-06-29 2014-01-02 Hewlett-Packard Development Company, L.P. Rule-based automated test data generation
US8904354B2 (en) * 2012-12-31 2014-12-02 Ca, Inc. Rule based syntax software test case generator
CN103559131B (en) * 2013-11-07 2017-04-26 中国科学院软件研究所 Effective layering error locating method
US10621064B2 (en) 2014-07-07 2020-04-14 Oracle International Corporation Proactive impact measurement of database changes on production systems
CN104063325B (en) * 2014-07-11 2016-10-26 电子科技大学 A kind of test exemple automation generating means for embedded software and method thereof
US10176085B2 (en) * 2015-06-24 2019-01-08 Tata Consultancy Services Limited Method and system for generating functional test cases for software systems
CN107153609B (en) * 2016-03-04 2020-12-22 创新先进技术有限公司 Automatic testing method and device
WO2018060777A1 (en) * 2016-09-29 2018-04-05 Yokogawa Electric Corporation Method and system for optimizing software testing
US10467130B2 (en) * 2017-01-20 2019-11-05 Wipro Limited Method of generating and transforming test data and a system therefor
US10545962B2 (en) 2017-01-25 2020-01-28 Salesforce.Com, Inc. Relational database instruction validation
US10650028B2 (en) * 2017-01-25 2020-05-12 Salesforce.Com, Inc. Relational database instruction validation
US10204034B2 (en) * 2017-04-06 2019-02-12 At&T Intellectual Property I, L.P. System and method for testing software applications in a software defined network
US11386058B2 (en) 2017-09-29 2022-07-12 Oracle International Corporation Rule-based autonomous database cloud service framework
US11327932B2 (en) 2017-09-30 2022-05-10 Oracle International Corporation Autonomous multitenant database cloud service framework
CN107766490A (en) * 2017-10-17 2018-03-06 链家网(北京)科技有限公司 Automatic operation method, system, equipment and the storage medium of SQL files
US20190188118A1 (en) * 2017-12-15 2019-06-20 Unisys Corporation System and method for generating database independent input test data
JP7044086B2 (en) 2019-03-15 2022-03-30 オムロン株式会社 Control systems, control methods, and control programs
US11036619B2 (en) 2019-06-06 2021-06-15 International Business Machines Corporation Bypassing execution of a module in real-time
US10915426B2 (en) 2019-06-06 2021-02-09 International Business Machines Corporation Intercepting and recording calls to a module in real-time
US10929126B2 (en) 2019-06-06 2021-02-23 International Business Machines Corporation Intercepting and replaying interactions with transactional and database environments
US11074069B2 (en) * 2019-06-06 2021-07-27 International Business Machines Corporation Replaying interactions with transactional and database environments with re-arrangement
US11016762B2 (en) 2019-06-06 2021-05-25 International Business Machines Corporation Determining caller of a module in real-time
US10909109B1 (en) * 2019-12-30 2021-02-02 Atlassi An Pty Ltd. Quality control test transactions for shared databases of a collaboration tool
CN113742200A (en) * 2020-05-28 2021-12-03 阿里巴巴集团控股有限公司 Test method, device and computer readable medium for database version upgrade
CN113535585A (en) * 2021-08-03 2021-10-22 广域铭岛数字科技有限公司 Test data generation method and system
US11899560B2 (en) * 2022-04-27 2024-02-13 Microsoft Technology Licensing, Llc Automatic correctness validation of database management systems

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5151697A (en) * 1990-10-15 1992-09-29 Board Of Regents Of The University Of Washington Data structure management tagging system
US5295256A (en) * 1990-12-14 1994-03-15 Racal-Datacom, Inc. Automatic storage of persistent objects in a relational schema
US5404295A (en) * 1990-08-16 1995-04-04 Katz; Boris Method and apparatus for utilizing annotations to facilitate computer retrieval of database material
US5434777A (en) * 1992-05-27 1995-07-18 Apple Computer, Inc. Method and apparatus for processing natural language
US5548770A (en) * 1993-02-25 1996-08-20 Data Parallel Systems, Inc. Method and apparatus for improving retrieval of data from a database
US5561421A (en) * 1994-07-28 1996-10-01 International Business Machines Corporation Access method data compression with system-built generic dictionaries
US5640550A (en) * 1994-04-15 1997-06-17 Coker; Drake Computer system for generating SQL statements from COBOL code
US5659737A (en) * 1995-08-01 1997-08-19 Oracle Corporation Methods and apparatus for data compression that preserves order by using failure greater than and failure less than tokens
US5761654A (en) * 1996-06-05 1998-06-02 Oracle Corporation Memory structure and method for tuning a database statement using a join-tree data structure representation, including selectivity factors, of a master table and detail table
US5764973A (en) * 1994-02-08 1998-06-09 Enterworks.Com, Inc. System for generating structured query language statements and integrating legacy systems
US5845285A (en) * 1997-01-07 1998-12-01 Klein; Laurence C. Computer system and method of data analysis
US5893102A (en) * 1996-12-06 1999-04-06 Unisys Corporation Textual database management, storage and retrieval system utilizing word-oriented, dictionary-based data compression/decompression
US6029043A (en) * 1998-01-29 2000-02-22 Ho; Chi Fai Computer-aided group-learning methods and systems

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5412806A (en) 1992-08-20 1995-05-02 Hewlett-Packard Company Calibration of logical cost formulae for queries in a heterogeneous DBMS using synthetic database
US5590319A (en) * 1993-12-15 1996-12-31 Information Builders, Inc. Query processor for parallel processing in homogenous and heterogenous databases
US5692107A (en) 1994-03-15 1997-11-25 Lockheed Missiles & Space Company, Inc. Method for generating predictive models in a computer system
US5584024A (en) * 1994-03-24 1996-12-10 Software Ag Interactive database query system and method for prohibiting the selection of semantically incorrect query parameters
CA2130065C (en) 1994-08-12 1999-03-02 Michael Joel Cincinatus Utilizing pseudotables as a method and mechanism for providing database monitor information
US5724570A (en) 1995-06-07 1998-03-03 Tandem Computers Incorporated Method and apparatus for a complete SQL subquery elimination process
US5701471A (en) 1995-07-05 1997-12-23 Sun Microsystems, Inc. System and method for testing multiple database management systems
US5732274A (en) 1995-11-08 1998-03-24 Electronic Data Systems Corporation Method for compilation using a database for target language independence
US5664173A (en) 1995-11-27 1997-09-02 Microsoft Corporation Method and apparatus for generating database queries from a meta-query pattern
US5950188A (en) 1996-11-14 1999-09-07 Sybase, Inc. Database system with methods for executing system-created internal SQL command statements
US5852818A (en) 1996-12-23 1998-12-22 Oracle Corporation Non-recursive method for parameter evaluation within an information management system
US6094649A (en) * 1997-12-22 2000-07-25 Partnet, Inc. Keyword searches of structured databases
US6581052B1 (en) * 1998-05-14 2003-06-17 Microsoft Corporation Test generator for database management systems
US6138112A (en) * 1998-05-14 2000-10-24 Microsoft Corporation Test generator for database management systems
US6631519B1 (en) * 2000-03-30 2003-10-07 Microsoft Corporation Automated schema and interface generation
KR100426307B1 (en) * 2001-12-12 2004-04-08 한국전자통신연구원 Apparatus and method accessing data by using single object access protocol-extended markup language

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5404295A (en) * 1990-08-16 1995-04-04 Katz; Boris Method and apparatus for utilizing annotations to facilitate computer retrieval of database material
US5151697A (en) * 1990-10-15 1992-09-29 Board Of Regents Of The University Of Washington Data structure management tagging system
US5295256A (en) * 1990-12-14 1994-03-15 Racal-Datacom, Inc. Automatic storage of persistent objects in a relational schema
US5434777A (en) * 1992-05-27 1995-07-18 Apple Computer, Inc. Method and apparatus for processing natural language
US5548770A (en) * 1993-02-25 1996-08-20 Data Parallel Systems, Inc. Method and apparatus for improving retrieval of data from a database
US5764973A (en) * 1994-02-08 1998-06-09 Enterworks.Com, Inc. System for generating structured query language statements and integrating legacy systems
US5640550A (en) * 1994-04-15 1997-06-17 Coker; Drake Computer system for generating SQL statements from COBOL code
US5561421A (en) * 1994-07-28 1996-10-01 International Business Machines Corporation Access method data compression with system-built generic dictionaries
US5659737A (en) * 1995-08-01 1997-08-19 Oracle Corporation Methods and apparatus for data compression that preserves order by using failure greater than and failure less than tokens
US5761654A (en) * 1996-06-05 1998-06-02 Oracle Corporation Memory structure and method for tuning a database statement using a join-tree data structure representation, including selectivity factors, of a master table and detail table
US5893102A (en) * 1996-12-06 1999-04-06 Unisys Corporation Textual database management, storage and retrieval system utilizing word-oriented, dictionary-based data compression/decompression
US5845285A (en) * 1997-01-07 1998-12-01 Klein; Laurence C. Computer system and method of data analysis
US6029043A (en) * 1998-01-29 2000-02-22 Ho; Chi Fai Computer-aided group-learning methods and systems

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9152630B1 (en) * 2011-10-26 2015-10-06 Intuit Inc. Modified database transaction scopes during software testing
US20140101097A1 (en) * 2012-10-09 2014-04-10 Sap Ag Template based database analyzer
US9311345B2 (en) * 2012-10-09 2016-04-12 Sap Se Template based database analyzer
US9015165B1 (en) * 2014-03-03 2015-04-21 Michael L. Hamm Text-SQL relational database

Also Published As

Publication number Publication date
US20030191774A1 (en) 2003-10-09
US6581052B1 (en) 2003-06-17
US7007007B2 (en) 2006-02-28
US20050097118A1 (en) 2005-05-05
US6826558B2 (en) 2004-11-30

Similar Documents

Publication Publication Date Title
US6826558B2 (en) Test generator for database management systems providing tight joins
US6138112A (en) Test generator for database management systems
He et al. Transform-data-by-example (TDE) an extensible search engine for data transformations
Khoussainova et al. Perfxplain: debugging mapreduce job performance
Pan et al. Guided test generation for database applications via synthesized database interactions
US11385889B2 (en) Inferring intra package and module dependencies
Signore et al. Reconstruction of er schema from database applications: a cognitive approach
US7831614B2 (en) System and method for generating SQL using templates
Nagy et al. Where was this SQL query executed? a static concept location approach
WO2005022409A2 (en) Information system development
Powell Oracle performance tuning for 10gR2
Khurana et al. Opaque Query Extraction
MySQL MySQL administrator's guide and language reference
US7058620B1 (en) Cross-platform subselect metadata extraction
Zhang et al. Duplicate-sensitivity Guided Transformation Synthesis for DBMS Correctness Bug Detection
Powell Oracle High Performance Tuning for 9i and 10g
Egger SQL in the Cloud
Anderson Modeling and analysis of SQL queries in PHP systems
Ježek et al. NWB Query engines: tools to search data stored in neurodata without borders format
Faroult et al. Refactoring SQL applications
Klamut Developing the SQL IonDB Query Language
Bhat et al. Small-Scale Relational Database Management System
Figueroa et al. A Brief Comparison of Two Enterprise-Class RDBMSs
Fallmann et al. Comparison of the Enterprise Functionalities of Open Source Database Management Systems
Pereira et al. Endowing NoSQL DBMS with SQL Features Through Standard Call Level Interfaces.

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034766/0001

Effective date: 20141014