US20130246287A1 - Method and apparatus for testing automated valuation models - Google Patents

Method and apparatus for testing automated valuation models Download PDF

Info

Publication number
US20130246287A1
US20130246287A1 US13/756,927 US201313756927A US2013246287A1 US 20130246287 A1 US20130246287 A1 US 20130246287A1 US 201313756927 A US201313756927 A US 201313756927A US 2013246287 A1 US2013246287 A1 US 2013246287A1
Authority
US
United States
Prior art keywords
data
automated valuation
automated
valuation
rankings
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/756,927
Inventor
Benjamin Graboske
Robert L. Walker
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CoreLogic Solutions LLC
Original Assignee
CoreLogic Solutions LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CoreLogic Solutions LLC filed Critical CoreLogic Solutions LLC
Priority to US13/756,927 priority Critical patent/US20130246287A1/en
Publication of US20130246287A1 publication Critical patent/US20130246287A1/en
Assigned to BANK OF AMERICA, N.A. reassignment BANK OF AMERICA, N.A. SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CORELOGIC SOLUTIONS, LLC
Assigned to CORELOGIC SOLUTIONS, LLC reassignment CORELOGIC SOLUTIONS, LLC RELEASE OF SECURITY INTEREST RECORDED AT 032798/0047 Assignors: BANK OF AMERICA, N.A.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0278Product appraisal
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q40/00Finance; Insurance; Tax strategies; Processing of corporate or income taxes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q40/00Finance; Insurance; Tax strategies; Processing of corporate or income taxes
    • G06Q40/12Accounting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/16Real estate

Definitions

  • the present invention relates to testing procedures to evaluation property valuations, and more specifically to a method of and apparatus for testing the accuracy of multiple automated valuation models.
  • Automated valuation models are increasingly being used in real estate as the first or sole indicator of property value.
  • the valuations provided by these automated valuation models are of varying accuracy.
  • the “confidence scores” often included along with any automated valuation produced are only indicators of a particular automated valuation model's internal evaluation that its own valuation is more or less accurate and often bear no relation to the internal evaluations of other automated valuation model valuations.
  • an automated valuation model may be tested for accuracy.
  • This method and apparatus further provides a means by which comparisons between various automated valuation models may be performed using standardized indications of accuracy.
  • the method and apparatus of this invention also provides data from which any user of automated valuation model data may perform calculations and comparisons of whatever kind using standardized data.
  • the method begins with the receipt of new sales data on a property.
  • data may be generated in several ways in alternative embodiments including appraisals, foreclosures, and even other automated valuations.
  • This new data is quickly passed on to the real time testing implementation to be described.
  • This new data is inputted in real-time, as it is received, into a database.
  • the data saved includes the property's sale (or other valuation) price and the property's geographic area.
  • the real time testing implementation requests an automated valuation of the new data's subject property. This request is forwarded to each automated valuation model to which the particular real time testing implementation has access. Each automated valuation model then returns a valuation without reference to the current sale of that subject property. It is unlikely that an automated valuation model has been updated to take this new property sale, appraisal, or other valuation into account. This is due to the real time nature of the invention. Because automated valuation models are not updated in real time, it would be highly unlikely that any one automated valuation model would have already received and incorporated the new data that has just been received by the method of this invention.
  • an implementation of this invention may aggregate automated valuation model valuation requests until the end of a day or until the end of any other predetermined period of time. Then, the implementation of this invention requests a valuation for each property for which new sale or reference data was generated since the last request from each automated valuation model being ranked.
  • the database of reference data and automated valuation model valuations is thereby continuously updated, but need not necessarily be in real time. This automated valuation model data will almost certainly not have been updated to include the new sales data because automated valuation models are usually only updated once a month or quarter.
  • the continuously updated database of automated valuations and of reference values provides a data set from which the data used to rank the accuracy of automated valuation models can be derived.
  • the database containing the sale price (or other new data) and each automated valuation model valuation of the subject property may then be used to create various indicators of the accuracy of each automated valuation model's property valuations. These indicators may be calculated for any geographic area, property type, economic tier, or any other characteristic of a property.
  • the data within the database is used to create several numerical indicators of the accuracy of an automated valuation model in a given geographic area.
  • the automated valuation models are then ranked, using ordinal numbers, according to their accuracy. These numerical indicators are then weighed, depending upon their usefulness as indicators of overall accuracy. Alternative weightings may be given to particular indicators. In another embodiment, the weighting of the various indicators of accuracy may be done using an equation.
  • the rankings may be calculated based on price ranges or taking into account the cost to request each automated valuation model valuation, such that a ranking may be generated for the automated valuation model of the best accuracy for the cost or the automated valuation model that is best in a particular price range or geographic area.
  • secondary rankings may be made as to accuracy excluding data referenced by the primary (or most accurate) automated valuation model overall.
  • the secondary rankings may be further subdivided to exclude only the data referenced by the primary (or most accurate) automated valuation model in a particular geographic area, economic tier or property type.
  • the rankings of accuracy in a particular geographic area, economic tier or property type may then be passed on to the users of automated valuations models.
  • the rankings may be designed to cater to a particular client, incorporating other interests, such as least-expensive accurate automated valuation model.
  • the data itself indicating the accuracy of each automated valuation model may simply be passed directly, either in total or piecemeal, to a user of automated valuation models. This data may be used to perform whatever calculations the particular user desires in order to determine what automated valuation model is best for the particular user's needs.
  • the data may be passed along to an automated valuation model user along with the ranking data as calculated in the preferred embodiment, such that the automated valuation model user may ensure the accuracy of the ranking database information and to perform their own calculations and weightings.
  • FIG. 1 is a data structure used to implement the method and apparatus of the invention.
  • FIG. 2 is a more detailed depiction of specific portions of the data structure depicted in FIG. 1 along with external components used in the method of this invention.
  • FIG. 3 is a flowchart depicting the steps involved in the data collection process for each property.
  • FIG. 4 is a flowchart depicting the steps involved in the periodic ranking of automated valuation model accuracy and updating of databases.
  • FIG. 5 a is a table depicting an example of a primary ranking of several automated valuation models using a predetermined criterion.
  • FIG. 5 b is a table depicting an example of a secondary ranking of the remaining automated valuation models using the predetermined criterion.
  • FIG. 6 a is a table depicting a summary of the primary weighted rankings of several automated valuation models using the data obtained from the rankings show in FIG. 5 a.
  • FIG. 6 b is a table depicting a summary of the secondary weighted rankings of several automated valuation models using the data obtained from the rankings shown in FIG. 5 b.
  • FIG. 7 a is a table depicting an example alternative embodiment of a primary ranking of several automated valuation models using an alternative predetermined criterion.
  • FIG. 7 b is a table depicting an example alternative embodiment of a secondary ranking of the remaining automated valuation models using the predetermined criterion.
  • FIG. 8 a is a table depicting a summary of the primary weighted rankings of several automated valuation models using the data obtained from the rankings shown in FIG. 7 a.
  • FIG. 8 b is a table depicting a summary of the secondary weighted rankings of several automated valuation models using the data obtained from the rankings shown in FIG. 7 b.
  • the present invention provides a method of and apparatus for real time testing of automated valuation models (“AVMs” or singularly “AVM”).
  • AVMs automated valuation models
  • FIG. 1 an example data structure for use in implementing this method is depicted.
  • element 90 Depicted in element 90 is an example automated valuation model testing data structure (“data structure”).
  • data structure 90 may be altered, and some may even be eliminated, without diminishing the scope of the invention.
  • This data structure 90 is one means by which to implement the method of the invention. Many alternative embodiments of the present invention may be used which accomplish substantially the same result.
  • the control processor 100 is used as an internal control mechanism. If implemented in software, for example, it is the main function of the program, used in calling each subsidiary function. If implemented in hardware, it is the chip in which each of the other functions is embodied.
  • the automation processor 102 is used in automating the process by which the new sale data or other new data is incorporated into the real time testing database. It also automates the process by which automated valuation data is requested from each automated valuation model being tested and the process by which periodic calculations and rankings of the various automated valuation models are performed.
  • the categorization processor 104 is used to categorize property according to its geographic area, economic tier or property type.
  • Economic tier refers to a particular valuation range, price range, valuation tier or price tier of which a property may be a part. Tiers may be divided into quartiles, made up of upper tier, two middle tiers and a lower tier. Tiers may be calculated as a percentile of valuation or price in relation to other properties in a large or small geographic area or a property type. Economic tier may be determined using any price or valuation related data concerning a particular property that would allow categorization of a property with respect to another property.
  • Property type refers to a property type such as a single-family residence, condominium or any other means of categorizing properties. Alternatively, property type could also be a categorization based on the number of bedrooms and bathrooms in a particular property. Any method of grouping and differentiating properties based upon their individual characteristics could be considered a property type.
  • the calculation processor 106 is used to perform the calculations associated with deriving indicators of accuracy for each automated valuation model and it is also used in using those indicators and the various weighting factors to order the automated valuation models for ranking.
  • the comparison processor 107 is used in comparing the reference values with the automated valuations.
  • Temporary data storage 108 is used to temporarily store data as it is being used. In a standard personal computer, temporary data storage is random, access memory (RAM). It is used to store temporary data, such as partial calculations of the indicators of accuracy or ranking data before it is stored into the AVM accuracy database 120 .
  • the input and output connectors 110 are used as a buffer between the data 90 and external data and output sources such as the new data source 114 .
  • the input and output connectors use standard communications protocols when able to do so and use proprietary communications protocols when necessary to communicate with a particular data or output source.
  • the elements connected to the data structure 90 need not actually be separate from the data structure 90 . They are depicted as connected in this example database, but some or all may be included within the data structure 90 itself, thus eliminating the need for such a buffer for that input or output source.
  • the new data source in the preferred embodiment is a continually updated source of real estate new sales data. It need not necessarily be one source.
  • the new data source 114 may be representative of many sources of sales data.
  • the reference data derived from the new data source 114 is an actual sale of a property. In alternative embodiments, reference data from the new data source 114 may be derived by using appraisals or from other automated valuation model valuations.
  • the price data and valuation database 116 is also connected to the data structure 90 through the input and output connectors 110 . This is the database wherein all the new sale prices and each AVM valuation for a subject property are stored.
  • this price data is derived from new sales, but as stated above, this data may also be from appraisals or from AVM valuations.
  • This price data and valuation database 116 is depicted as a single database. In alternative embodiments, it may be multiple databases, one for each automated valuation model being tested and one for new sale data. Alternative categorization into multiple databases is possible in alternative embodiments of this invention.
  • the additional input and output resources 118 are also connected to the data structure 90 of this invention using the input and output connectors.
  • This element is representative of any additional input or output necessary or useful in the functioning of the real time testing data structure 90 .
  • These additional input and output resources may include a monitor, a keyboard, a mouse, a network connection, a modem connection, a form of wireless connection, a direct connection to another computer or network, an Internet connection for web access to a particular database or user interface, or a connection to other databases or data sources. These are only examples of potential input and output resources. Many others are potentially possible and useful, though they are not depicted here.
  • the AVM accuracy database 120 is also connected to the data structure 90 using the input and output connectors. This database need not necessarily be separate from the data structure 90 , but is depicted herein as being separate. This database stores the automated valuation model rankings, which in the preferred embodiment are updated in real time, of each automated valuation model being tested. This database is represented as being a single entity. It may also be divided and subdivided into accuracy databases for particular geographic areas, economic tiers, and property types. Individual AVM accuracy databases 120 may be created for any identifiable grouping of properties for which data is available.
  • the automated valuation model connector 112 is used to connect the data structure 90 implementation of this invention with the multiple automated valuation models being tested. It performs a function similar to that of the input and output connectors, specifically designed with multiple automated valuation model valuation requests and responses in mind.
  • the transactions between an automated valuation model and the invention such as AVM X valuation request 128 and AVM X valuation response 130 , are performed using extended markup language (XML). Often used in communications between different software programs or data structures, this language is used by the automated valuation model connector 112 to format the valuation request for each automated valuation model being tested.
  • XML extended markup language
  • Example automated valuation models AVM X 122 , AVM Y 124 , and AVM Z 126 are depicted as connected to the automated valuation model connector. Each of these AVMs may have a different format in which it expects automated valuation requests and responses to be forwarded and returned.
  • the automated valuation model connector 112 is designed to handle and format requests and responses for each of these AVMs. These three connected AVMs are only examples. Additional AVMs may be connected. In the preferred embodiment, all available AVMs are connected to enable the real time testing method to provide the most complete data and accuracy rankings for AVMs.
  • the first step in the method of this invention is to obtain new data or a reference value generated from a new data source 114 .
  • this is generally a new sale of a property, but may in alternative embodiments be an appraisal or even a particular automated valuation model valuation (for use in comparing other AVMs to the one used as the new sale data).
  • the new data is inputted through the input and output connectors 110 into the new data and valuation database 116 .
  • the second step involves requesting and receiving automated valuations from each of the automated valuation models being tested for each reference value.
  • This request may occur immediately subsequent to the reception of new automated valuation model data or at some time shortly thereafter. In the preferred embodiment, this occurs within one day to one week from the receipt of new sale or reference data.
  • example AVMs X 122 , Y 124 and Z 126 are depicted.
  • the various indicators of accuracy are calculated as percentages. These indicators are: absolute mean error, absolute median variance to the reference value (usually the new sale price), mean squared error, automated valuation model accuracy (defined as the percentage of valuations within a predetermined percentage of the reference value, such as 10% of the reference value), the percentage incidence of outlier values (defined as valuations that are beyond a predetermined percentage of the reference value, such as 25% more or less than the reference value), and hit rate percentage (defined as percentage of properties for which the AVM was able to return an evaluation).
  • These indicators of accuracy are calculated periodically or on command by a user based upon the aggregate data collected over time using the above process of requesting a valuation for a property for which new sale data or a reference value has been generated. Over time, this will provide an increasingly more accurate picture of a particular automated valuation model's accuracy overall and in a particular geographic area, economic tier, or property type. Additionally, these indicators of accuracy may be calculated for particular time frames, for example, if some improvements have been made recently to a particular automated valuation model, then the indicators of accuracy and subsequent rankings dependant upon them may be calculated for the time since the improvements were made.
  • weighting factors are as follows:
  • hit rate is multiplied by 0.05 or 5%
  • outlier percentage is multiplied by 0.30 or 30%
  • n′′ is the number of properties for which a valuation was returned and n is the total number of properties.
  • s n is the sale price or new data price for property n
  • r n is the reference value price for property n
  • n is the number of properties used in the calculation.
  • n is the number of properties used and i is the list of those errors listed in ascending order by absolute value.
  • s n is the sale price or new data price for property n
  • r n is the reference value price for property n
  • n is the number of properties used in the calculation.
  • s n is the sale price or new data price for property n
  • r n is the reference value price for property n
  • n is the number of properties used in the calculation.
  • n′ is the number of properties whose valuation was within a predetermined percentage of the reference value, such as 10% of the reference value, and where n is the total number of properties. This percentage may be greater or smaller without affecting the method of this invention.
  • Percentage incidence of outlier values (o) is calculated using the following equation:
  • o′ is the number of properties whose valuation was 25% more or less than the reference value and n is the total number of properties.
  • the percentage that is considered to be an outlier value may be greater or smaller without affecting the method of this invention.
  • each of the above calculations results in a decimal number that is representative of a percentage value.
  • a large percentage may indicate accuracy or may indicate inaccuracy.
  • a large percentage indicates accuracy.
  • the highest hit rate percentage is ranked number one in a particular geographic area, economic tier, or property type.
  • a lower percentage indicates more accuracy, stating that the average percentage of value that this particular automated valuation model “misses” the target. For this indicator and in this embodiment, the lowest percentage is ranked highest.
  • each automated valuation model is ordinally ranked as compared with every other automated valuation model.
  • these rankings are then multiplied, for each of the automated valuation models in each geographic area, economic tier, or property type, by the associated weighting factor.
  • Each of these multiplied values are added together and multiplied by a scaling factor, such as ten, which results in an “accuracy score.”
  • the lowest accuracy score in a particular geographic area, economic tier or property type is the most accurate automated valuation model.
  • a secondary ranking may take place.
  • the same reference values usually sale prices, but sometimes alternative new price data, are considered.
  • the computation of rankings is performed using data for which the primary ranked AVM did not return an evaluation.
  • the hit rate is calculated for the primary ranking using all reference values and automated valuation model valuations for each subject property.
  • the secondary hit rate is calculated using only reference values and automated valuation model valuations for subject properties for which the primary automated valuation model did not return a valuation.
  • the absolute mean error would only be calculated for properties for which there was no valuation returned by the primary ranked automated valuation model in that geographic area, economic tier, or property type.
  • the purpose of this secondary ranking is to help enable automated valuation model users to determine which automated valuation model to use second, if the primary ranked automated valuation model is unable to value a particular subject property.
  • any external client databases 136 are updated. This connection is made through the input and output connectors 110 .
  • the external client database 136 is a database of a client to which the new rankings and accuracy indicators and weights are sent. Additionally, the new data generated by the calculation processor 106 and the comparison processor 107 is sent through the input and output connectors 110 and appended to the AVM accuracy database 120 .
  • the rankings, reference values and automated valuations may be added to a new AVM accuracy database 120 for only a particular time period, such as a three-month period. In either database update, the database 120 is updated to reflect changes in the rankings of automated valuation models and the numerical changes to the various indicators of accuracy. Additional data users 138 may also be updated through the input and output connectors 110 .
  • the new data source 114 will likely be used by relevant public agencies to update public record databases 132 .
  • various automated valuation model venders periodically use this new data, generally on a monthly or quarterly basis, to update the databases of each of the automated valuation models 134 .
  • FIG. 3 depicts the steps involved in the data collection process of the present invention.
  • the first step in this process is to gather or generate new property value indication data 140 .
  • This entails the reception of new sale data from publicly available records.
  • the data is appended to the database as quickly as it is received.
  • this new data is received and stored 142 in the new data and valuation database 116 .
  • valuation requests are made to each automated valuation model to be tested 144 . This request need not occur immediately after receipt of new sale data.
  • the new sale data is aggregated and all automated valuations are requested at the end of a day.
  • the first step is to calculate the accuracy indicators 148 .
  • Each of the indicators of accuracy are calculated for each automated valuation model. This is done using the aggregate reference values and automated valuations for each subject property over a set period of time. This time period may be over the life of a particular automated valuation model or for using all data that the method of this invention has been able to store. Alternatively, these indicators of accuracy may be calculated for only a smaller subset of all of the data available because of the method of this invention.
  • Each of the indicators of accuracy is a percentage or an average percentage of a plurality of valuations that may be used to show, in general, the accuracy of each of the tested automated valuation models.
  • each AVM is ranked in each indicator of accuracy and in each category-geographic area, economic tier, or property type-by ordinal numbers, then the ordinal number for each AVM's ranking is multiplied by its corresponding weighting factor, and the products of these multiplications are summed 150 .
  • the corresponding sum of the products is then used to rank the automated valuation models 152 in the embodiments described.
  • the lowest summed number is the most accurate automated valuation model, and the next lowest summed number is the next most accurate.
  • an overall accuracy indicating score is created using an additional formula, rather than ranking each automated valuation model compared to another.
  • each of the accuracy indicators is calculated in much the same way, but they are weighted using a formula or equation.
  • the highest calculated number is the most accurate automated valuation model, and the next highest calculated number is the next most accurate.
  • the automated valuation models are then ranked according to which has the highest accuracy score. This alternative embodiment will be described further in discussing FIGS. 7 a - 8 b.
  • the method of this invention updates the AVM accuracy database 154 .
  • This database then contains the most up to-date rankings and the indicators that led to those rankings.
  • FIG. 5 a the contents of a primary ranking database are depicted.
  • the various indicators of accuracy are depicted in column 155 .
  • Hit rate percentage 156 refers to the percentage of reference properties for which each automated valuation model was able to return a valuation.
  • Median error 158 refers to the percentage error at the middle range of the ordered “list” of errors with reference to the reference value.
  • Absolute mean error 160 refers to the percentage average of the absolute value of the list of errors in valuation.
  • Mean squared error 162 is the percentage average in relation to the reference value of the squares of the “list” of errors in valuation.
  • AVM accuracy 164 refers to the percentage of automated valuations within plus or minus 10% of the reference value.
  • Outlier percentage 166 refers to the percentage of automated valuations that are more than plus or minus 25% of the reference value.
  • the rank for each evaluated AVM based on factors 156 through 166 for a predetermined geographic area is in column 170 .
  • the rankings for three AVMs, X, Y, and Z, are shown in FIG. 5 a for three states, Florida, California, and New Jersey. A lower ranking number indicates higher accuracy. For example, the rankings for three AVMs for Florida are shown, AVM X 172 , AVM Y 174 , and AVM Z 176 .
  • the various factors are compared and then ranked. For example in element 178 , the hit rate percentage of AVM X in Florida is 93.6%.
  • the hit rate percentage is 85.1%.
  • the percentage is 67.8%. Because the higher the hit rate, that is the number of properties for which a response was generated by the automated valuation model, the higher the accuracy, a larger percentage for this factor is given a higher rank. Therefore, AVM X is ranked # 1 in element 184 , AVM Y is ranked # 2 in element 186 and AVM Z is ranked # 3 in element 188 .
  • AVM Accuracy depicted in element 164 , is returned as a percentage. This represents, as stated above, the percentage of values that were within plus or minus 10% of the actual price or other reference value.
  • the percentage for AVM X of accuracy was 49.2%, depicted in element 190 .
  • the same percentage for AVM Y was 53.4%, depicted in element 192 .
  • the same percentage for AVM Z was 50.5%, depicted in element 194 . This results in a ranking of # 3 for AVM X in element 196 , of # 1 for AVM Y in element 198 and a ranking of # 2 for AVM Z in element 200 .
  • weighting factors are as follows:
  • hit rate is multiplied by 0.05 or 5%
  • outlier percentage is multiplied by 0.30 or 30%
  • the score for AVM Y in Florida is 13.5, as depicted in element 204 .
  • the score for AVM Z in Florida is 23.5, as depicted in element 206 .
  • a secondary ranking provides a ranking of the automated valuation models for properties for which the automated valuation model ranked first in the primary ranking was unable to return a valuation. This ranking is therefore useful in determining which automated valuation model is a good “second choice.”
  • This database represents most of the same calculations as those of the primary raking, minus the data that was already calculated using the primary ranked automated valuation model, that is the automated valuation model that ranked number 1 overall in a particular category or geographic area. Similar factors are considered along the left column. These are hit rate percentage 212 , median error 214 , absolute mean error 216 , mean squared error 218 , AVM accuracy 220 and outlier percentage 222 .
  • This secondary ranking is used to determine which automated valuation model to use after the primary ranked model. There is an importance in providing any valuation after the primary ranked automated valuation model was unable to provide one.
  • the hit rate percentage of 70.3% is depicted, as shown in element 213 .
  • the score 224 is also depicted.
  • the rank 226 or 230 the automated valuation model being tested, such as AVM X—Florida 228 or AVM Z—Florida 232 are depicted.
  • the weighting factors are as follows:
  • hit rate is multiplied by 0.05 or 5%
  • outlier percentage is multiplied by 0.30 or 30%
  • the secondary ranking score for AVM Z in Florida is 20, as depicted in element 236 .
  • FIG. 6 a a summary of the example primary weighted ranking is depicted. This is a summary form of the results tabulated in FIG. 5 a .
  • each automated valuation model 238 is depicted.
  • the Florida ranking 240 is shown at the top.
  • the score 242 from FIG. 5 a and rankings, based upon the score from FIG. 5 a are depicted.
  • the automated valuation model with the lowest score in a particular area is the most accurate in that area, according to one embodiment of this invention. For example, AVM Y 246 with a score of 13.5, depicted in element 248 is ranked # 1 in element 250 .
  • FIG. 6 b an example summary secondary weighted ranking is depicted.
  • This table is similar to the table depicted in FIG. 6 a , but is based on the data contained in FIG. 5 b . Therefore, this data excludes the valuations that were theoretically done using the primary ranked automated valuation model that is the most accurate one in a particular geographic area.
  • This weighted ranking displays in the tabular form, automated valuation model 264 , score 266 , and rank 268 .
  • the Florida ranking is depicted in element 270 . In Florida, AVM X 272 had a secondary score of 10, depicted in element 274 and a ranking of # 1 , depicted in element 276 . All of the remaining automated valuation models are ranked in this secondary ranking.
  • FIG. 7 a the results of an alternative embodiment of the method used to implement the primary ranking of this invention is depicted.
  • a formula is used to provide the relative weightings of the individual indicators of accuracy.
  • the elements depicted in FIG. 7 a are much the same as those depicted in FIG. 5 a .
  • the indicators of accuracy hit rate percentage 278 through outlier percentage 280 are the same.
  • the values for hit rate percentage 278 in each automated valuation model are the same as in FIG. 5 a .
  • the hit rate percentage for AVM X in Florida is 93.6% as depicted in element 288 .
  • the outlier percentage, depicted in element 290 is 12.7%.
  • each of the same automated valuation models are depicted; AVM X in 282 , AVM Y in 284 and AVM Z in 286 .
  • the formulas used to calculate each of the percentage values are the same. However, an overall accuracy indicating score 292 is created using an additional formula, rather than ranking each automated valuation model's accuracy indicators compared to another. The following is the score computation equation where each percentage is represented as a decimal:
  • Score f *(( h+c ) ⁇ ( m+a+s+o ))
  • This equation is used to calculate each of the scores for the automated valuation models in each geographic area, economic tier or property type.
  • a multiplier factor of 500 is used in the invention to accentuate the differences in accuracy between the automated valuation model valuations.
  • a higher score represents a more accurate automated valuation in a particular geographic area, economic tier or property type.
  • AVM X is the most accurate with a score of 585, depicted in element 294 .
  • AVM Y is the next-most accurate with a score of 566, depicted in element 296 and
  • AVM Z is the least accurate with a score of 455, depicted in element 298 .
  • the secondary ranking is useful in finding a “second choice” automated valuation model.
  • the valuations performed by the first-ranked automated valuation model in the primary ranking are removed.
  • the “overall” primary automated valuation model was removed. This means that because AVM Y was the most accurate overall, it was “chosen” as the primary automated valuation model for each region.
  • the primary automated valuation model may be chosen for a particular geographic region, economic tier or property type. In that embodiment, for example, in Florida, AVM X would have been the primary automated valuation model because it was more accurate in that geographic area. So, it would be excluded from the secondary ranking.
  • Score f *(( h+c ) ⁇ ( m+a+s+o ))
  • FIG. 8 a an example summary primary weighted ranking of the alternative embodiment is depicted.
  • This is a summary form of the data presented in FIG. 7 a depicting only the scores in each location and the relative ranking. Similar to FIG. 6 a above, the state for which the rankings are being done is depicted as Florida in element 320 .
  • the column for automated valuation model 322 includes AVM X 328 , AVM Y 330 and AVM Z 332 .
  • the respective scores for each of these automated valuation models are 585 , depicted in element 334 ; 566 , depicted in element 336 ; and 455 , depicted in element 338 .
  • AVM X 328 with its score of 585 in Florida is ranked number one, as depicted in element 340 .
  • the overall rankings depicted below are a result of the addition of each of the scores in each area.
  • AVM Y 342 with its overall score of 1861, depicted in element 344 is ranked number one 346 overall.
  • the overall ranking may be determined by dividing the geographic area score by the number of properties in that area for which automated valuations were provided and then adding the numbers together. This would provide a ranking weighted as to the percentage of homes in a particular market or its relative importance in use of automated valuation data.
  • FIG. 8 b a secondary weighted ranking is depicted. This is a summary table of the data contained in FIG. 7 b . This table is useful in comparing directly secondary rankings.
  • the automated valuation model being ranked is depicted in the column denoted by element 350 .
  • the Florida secondary ranking for example, is depicted in element 348 .
  • AVM X in Florida for example, depicted in element 354 , received a rank of one, as depicted in element 356 .
  • the column depicted in element 352 denotes the rank.
  • AVM X 358 received the number one ranking 360 with a score of 1079, depicted in element 362 .

Abstract

A method and apparatus for real time testing of automated valuation using various indicators of accuracy. These indicators are then weighted according to their value as indicators of accuracy using individualized weighting factors or an equation. A ranking is then computed based upon the factors and their weights. This method is preformed continuously, so as to achieve real-time or periodically updated automated valuation model accuracy rankings.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 13/182,289, filed Jul. 13, 2011, titled “METHOD AND APPARATUS FOR TESTING AUTOMATED VALUATION MODELS,” which is a division of U.S. patent application Ser. No. 11/007,750, filed Dec. 8, 2004, titled “METHOD AND APPARATUS FOR TESTING AUTOMATED VALUATION MODELS,” now U.S. Pat. No. 8,001,024, each of which is hereby incorporated by reference herein in its entirety.
  • BACKGROUND
  • 1. Field of the Invention
  • The present invention relates to testing procedures to evaluation property valuations, and more specifically to a method of and apparatus for testing the accuracy of multiple automated valuation models.
  • 2. Background of the Invention
  • Automated valuation models are increasingly being used in real estate as the first or sole indicator of property value. The valuations provided by these automated valuation models are of varying accuracy. Currently, there exists no standardized means by which to compare the accuracy of one automated valuation model to that of another. The “confidence scores” often included along with any automated valuation produced are only indicators of a particular automated valuation model's internal evaluation that its own valuation is more or less accurate and often bear no relation to the internal evaluations of other automated valuation model valuations.
  • Therefore, a means by which these automated valuation models' accuracy at providing valuations would prove very useful. Providing a means to gauge the accuracy of one automated valuation model against that of another enables the automated valuation consumer to choose the preferable automated valuation model on which to rely. Generally speaking, automated valuation models may be more or less accurate depending on the geographic area or price range of the property. Therefore, a means of testing the accuracy of automated valuation models that is capable of categorizing valuations by geographic area, economic tier, or property type would also be useful. There currently exists no empirically based ranking system for automated valuation model accuracy. Further, there currently exists no accurate and up-to-date data upon which to base such a ranking system.
  • It is therefore an object of this invention to provide a means by which automated valuation models may be compared against each other for accuracy. It is a further object of this invention to provide this means in particular sub-divisions such as geographic areas, price ranges, price tiers, and property types. It is a further object of this invention to provide a useful data set for comparing these automated valuation models and for evaluating the rankings computed using this method.
  • These and other objectives of the present invention will become apparent from the following description of the invention.
  • SUMMARY OF THE INVENTION
  • According to the present invention, a method and apparatus are described whereby an automated valuation model may be tested for accuracy. This method and apparatus further provides a means by which comparisons between various automated valuation models may be performed using standardized indications of accuracy. The method and apparatus of this invention also provides data from which any user of automated valuation model data may perform calculations and comparisons of whatever kind using standardized data.
  • In the preferred embodiment of the invention, the method begins with the receipt of new sales data on a property. However, data may be generated in several ways in alternative embodiments including appraisals, foreclosures, and even other automated valuations. Once this new data is generated, it is quickly passed on to the real time testing implementation to be described. This new data is inputted in real-time, as it is received, into a database. The data saved includes the property's sale (or other valuation) price and the property's geographic area.
  • Immediately, the real time testing implementation requests an automated valuation of the new data's subject property. This request is forwarded to each automated valuation model to which the particular real time testing implementation has access. Each automated valuation model then returns a valuation without reference to the current sale of that subject property. It is unlikely that an automated valuation model has been updated to take this new property sale, appraisal, or other valuation into account. This is due to the real time nature of the invention. Because automated valuation models are not updated in real time, it would be highly unlikely that any one automated valuation model would have already received and incorporated the new data that has just been received by the method of this invention.
  • Alternatively, an implementation of this invention may aggregate automated valuation model valuation requests until the end of a day or until the end of any other predetermined period of time. Then, the implementation of this invention requests a valuation for each property for which new sale or reference data was generated since the last request from each automated valuation model being ranked. The database of reference data and automated valuation model valuations is thereby continuously updated, but need not necessarily be in real time. This automated valuation model data will almost certainly not have been updated to include the new sales data because automated valuation models are usually only updated once a month or quarter. Whether valuation requests are done in real time (immediately upon receipt of new sale data) or shortly thereafter (at the end of a predetermined time period), the continuously updated database of automated valuations and of reference values provides a data set from which the data used to rank the accuracy of automated valuation models can be derived.
  • The database containing the sale price (or other new data) and each automated valuation model valuation of the subject property may then be used to create various indicators of the accuracy of each automated valuation model's property valuations. These indicators may be calculated for any geographic area, property type, economic tier, or any other characteristic of a property. In one embodiment of the invention, the data within the database is used to create several numerical indicators of the accuracy of an automated valuation model in a given geographic area. The automated valuation models are then ranked, using ordinal numbers, according to their accuracy. These numerical indicators are then weighed, depending upon their usefulness as indicators of overall accuracy. Alternative weightings may be given to particular indicators. In another embodiment, the weighting of the various indicators of accuracy may be done using an equation. Then, the rankings may be calculated based on price ranges or taking into account the cost to request each automated valuation model valuation, such that a ranking may be generated for the automated valuation model of the best accuracy for the cost or the automated valuation model that is best in a particular price range or geographic area. Additionally, secondary rankings may be made as to accuracy excluding data referenced by the primary (or most accurate) automated valuation model overall. Alternatively, the secondary rankings may be further subdivided to exclude only the data referenced by the primary (or most accurate) automated valuation model in a particular geographic area, economic tier or property type.
  • The rankings of accuracy in a particular geographic area, economic tier or property type may then be passed on to the users of automated valuations models. The rankings may be designed to cater to a particular client, incorporating other interests, such as least-expensive accurate automated valuation model. Additionally, the data itself indicating the accuracy of each automated valuation model may simply be passed directly, either in total or piecemeal, to a user of automated valuation models. This data may be used to perform whatever calculations the particular user desires in order to determine what automated valuation model is best for the particular user's needs. Alternatively, the data may be passed along to an automated valuation model user along with the ranking data as calculated in the preferred embodiment, such that the automated valuation model user may ensure the accuracy of the ranking database information and to perform their own calculations and weightings.
  • Further features and advantages of the present invention will be appreciated by reviewing the following drawings and detailed description of the invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a data structure used to implement the method and apparatus of the invention.
  • FIG. 2 is a more detailed depiction of specific portions of the data structure depicted in FIG. 1 along with external components used in the method of this invention.
  • FIG. 3 is a flowchart depicting the steps involved in the data collection process for each property.
  • FIG. 4 is a flowchart depicting the steps involved in the periodic ranking of automated valuation model accuracy and updating of databases.
  • FIG. 5 a is a table depicting an example of a primary ranking of several automated valuation models using a predetermined criterion.
  • FIG. 5 b is a table depicting an example of a secondary ranking of the remaining automated valuation models using the predetermined criterion.
  • FIG. 6 a is a table depicting a summary of the primary weighted rankings of several automated valuation models using the data obtained from the rankings show in FIG. 5 a.
  • FIG. 6 b is a table depicting a summary of the secondary weighted rankings of several automated valuation models using the data obtained from the rankings shown in FIG. 5 b.
  • FIG. 7 a is a table depicting an example alternative embodiment of a primary ranking of several automated valuation models using an alternative predetermined criterion.
  • FIG. 7 b is a table depicting an example alternative embodiment of a secondary ranking of the remaining automated valuation models using the predetermined criterion.
  • FIG. 8 a is a table depicting a summary of the primary weighted rankings of several automated valuation models using the data obtained from the rankings shown in FIG. 7 a.
  • FIG. 8 b is a table depicting a summary of the secondary weighted rankings of several automated valuation models using the data obtained from the rankings shown in FIG. 7 b.
  • DETAILED DESCRIPTION OF THE INVENTION
  • The present invention provides a method of and apparatus for real time testing of automated valuation models (“AVMs” or singularly “AVM”). Referring first to FIG. 1, an example data structure for use in implementing this method is depicted. Depicted in element 90 is an example automated valuation model testing data structure (“data structure”). The elements in this example data structure 90 may be altered, and some may even be eliminated, without diminishing the scope of the invention. This data structure 90 is one means by which to implement the method of the invention. Many alternative embodiments of the present invention may be used which accomplish substantially the same result.
  • The following components each may be implemented in software, in hardware, or a combination of both, so long as the method of this invention is carried out by the particular implementation. The control processor 100 is used as an internal control mechanism. If implemented in software, for example, it is the main function of the program, used in calling each subsidiary function. If implemented in hardware, it is the chip in which each of the other functions is embodied. The automation processor 102 is used in automating the process by which the new sale data or other new data is incorporated into the real time testing database. It also automates the process by which automated valuation data is requested from each automated valuation model being tested and the process by which periodic calculations and rankings of the various automated valuation models are performed. The categorization processor 104 is used to categorize property according to its geographic area, economic tier or property type. Economic tier refers to a particular valuation range, price range, valuation tier or price tier of which a property may be a part. Tiers may be divided into quartiles, made up of upper tier, two middle tiers and a lower tier. Tiers may be calculated as a percentile of valuation or price in relation to other properties in a large or small geographic area or a property type. Economic tier may be determined using any price or valuation related data concerning a particular property that would allow categorization of a property with respect to another property.
  • Property type refers to a property type such as a single-family residence, condominium or any other means of categorizing properties. Alternatively, property type could also be a categorization based on the number of bedrooms and bathrooms in a particular property. Any method of grouping and differentiating properties based upon their individual characteristics could be considered a property type.
  • The calculation processor 106 is used to perform the calculations associated with deriving indicators of accuracy for each automated valuation model and it is also used in using those indicators and the various weighting factors to order the automated valuation models for ranking. The comparison processor 107 is used in comparing the reference values with the automated valuations. Temporary data storage 108 is used to temporarily store data as it is being used. In a standard personal computer, temporary data storage is random, access memory (RAM). It is used to store temporary data, such as partial calculations of the indicators of accuracy or ranking data before it is stored into the AVM accuracy database 120. The input and output connectors 110 are used as a buffer between the data 90 and external data and output sources such as the new data source 114.
  • The input and output connectors use standard communications protocols when able to do so and use proprietary communications protocols when necessary to communicate with a particular data or output source. The elements connected to the data structure 90 need not actually be separate from the data structure 90. They are depicted as connected in this example database, but some or all may be included within the data structure 90 itself, thus eliminating the need for such a buffer for that input or output source.
  • Data pertaining to a new property sale is provided as a new data source 114. The new data source in the preferred embodiment is a continually updated source of real estate new sales data. It need not necessarily be one source. The new data source 114 may be representative of many sources of sales data. In the preferred embodiment, the reference data derived from the new data source 114 is an actual sale of a property. In alternative embodiments, reference data from the new data source 114 may be derived by using appraisals or from other automated valuation model valuations. The price data and valuation database 116 is also connected to the data structure 90 through the input and output connectors 110. This is the database wherein all the new sale prices and each AVM valuation for a subject property are stored. In the preferred embodiment, this price data is derived from new sales, but as stated above, this data may also be from appraisals or from AVM valuations. This price data and valuation database 116 is depicted as a single database. In alternative embodiments, it may be multiple databases, one for each automated valuation model being tested and one for new sale data. Alternative categorization into multiple databases is possible in alternative embodiments of this invention.
  • The additional input and output resources 118 are also connected to the data structure 90 of this invention using the input and output connectors. This element is representative of any additional input or output necessary or useful in the functioning of the real time testing data structure 90. These additional input and output resources may include a monitor, a keyboard, a mouse, a network connection, a modem connection, a form of wireless connection, a direct connection to another computer or network, an Internet connection for web access to a particular database or user interface, or a connection to other databases or data sources. These are only examples of potential input and output resources. Many others are potentially possible and useful, though they are not depicted here.
  • The AVM accuracy database 120 is also connected to the data structure 90 using the input and output connectors. This database need not necessarily be separate from the data structure 90, but is depicted herein as being separate. This database stores the automated valuation model rankings, which in the preferred embodiment are updated in real time, of each automated valuation model being tested. This database is represented as being a single entity. It may also be divided and subdivided into accuracy databases for particular geographic areas, economic tiers, and property types. Individual AVM accuracy databases 120 may be created for any identifiable grouping of properties for which data is available.
  • The automated valuation model connector 112 is used to connect the data structure 90 implementation of this invention with the multiple automated valuation models being tested. It performs a function similar to that of the input and output connectors, specifically designed with multiple automated valuation model valuation requests and responses in mind. In the preferred embodiment, the transactions between an automated valuation model and the invention, such as AVM X valuation request 128 and AVM X valuation response 130, are performed using extended markup language (XML). Often used in communications between different software programs or data structures, this language is used by the automated valuation model connector 112 to format the valuation request for each automated valuation model being tested. This XML request is then sent to the particular automated valuation model being tested, the AVM then provides its response 130 in XML to the automated valuation model connector 112 which then translates it back into a form readable by the rest of the automated valuation model testing data structure 90. Example automated valuation models AVM X 122, AVM Y 124, and AVM Z 126 are depicted as connected to the automated valuation model connector. Each of these AVMs may have a different format in which it expects automated valuation requests and responses to be forwarded and returned. The automated valuation model connector 112 is designed to handle and format requests and responses for each of these AVMs. These three connected AVMs are only examples. Additional AVMs may be connected. In the preferred embodiment, all available AVMs are connected to enable the real time testing method to provide the most complete data and accuracy rankings for AVMs.
  • Referring now to FIG. 2, specific portions of the data structure along with external elements are depicted. This depiction contains several of the elements depicted in FIG. 1, but includes the more direct ways in which the components of this method are related. The first step in the method of this invention is to obtain new data or a reference value generated from a new data source 114. As stated above, this is generally a new sale of a property, but may in alternative embodiments be an appraisal or even a particular automated valuation model valuation (for use in comparing other AVMs to the one used as the new sale data). The new data is inputted through the input and output connectors 110 into the new data and valuation database 116. The second step involves requesting and receiving automated valuations from each of the automated valuation models being tested for each reference value. This request may occur immediately subsequent to the reception of new automated valuation model data or at some time shortly thereafter. In the preferred embodiment, this occurs within one day to one week from the receipt of new sale or reference data. In this figure, example AVMs X 122, Y 124 and Z 126 are depicted. Once the new data and the AVM valuations are received through the automated valuation model connector 112, they are sent through the input and output connectors 110 and stored in the new data and valuation database 116. The aggregate data is then used periodically or on command by a user by the calculation processor to calculate various indicators of accuracy for each AVM and to rank each AVM by the various indicators, to weight them according to weighting factors or a weighting equation, and then rank the automated valuation models accordingly.
  • The various indicators of accuracy are calculated as percentages. These indicators are: absolute mean error, absolute median variance to the reference value (usually the new sale price), mean squared error, automated valuation model accuracy (defined as the percentage of valuations within a predetermined percentage of the reference value, such as 10% of the reference value), the percentage incidence of outlier values (defined as valuations that are beyond a predetermined percentage of the reference value, such as 25% more or less than the reference value), and hit rate percentage (defined as percentage of properties for which the AVM was able to return an evaluation).
  • These indicators of accuracy are calculated periodically or on command by a user based upon the aggregate data collected over time using the above process of requesting a valuation for a property for which new sale data or a reference value has been generated. Over time, this will provide an increasingly more accurate picture of a particular automated valuation model's accuracy overall and in a particular geographic area, economic tier, or property type. Additionally, these indicators of accuracy may be calculated for particular time frames, for example, if some improvements have been made recently to a particular automated valuation model, then the indicators of accuracy and subsequent rankings dependant upon them may be calculated for the time since the improvements were made.
  • In the embodiment using weighting factors for each of the indicators of accuracy, the weighting associated with each of these factors is also important. In this embodiment, the weighting factors are as follows:
  • hit rate is multiplied by 0.05 or 5%
  • median error is multiplied by 0.1 or 10%
  • absolute mean error is multiplied by 0.1 or 10%
  • mean error squared is multiplied by 0.1 or 10%
  • automated valuation model accuracy is multiplied by 0.35 or 35%
  • outlier percentage is multiplied by 0.30 or 30%
  • Hit Rate Percentage
  • Hit rate percentage (h) is calculated using the following equation: h=n″/n
  • Where n″ is the number of properties for which a valuation was returned and n is the total number of properties.
  • Median Error
  • The individual property errors, used in calculating the median error to the reference values are calculated using the equation:

  • error=|s n −r n |/s n for every sale price (s) and reference value (r).
  • Where
  • sn is the sale price or new data price for property n,
  • rn is the reference value price for property n, and
  • n is the number of properties used in the calculation.
  • These errors are then ordered in ascending order such that the lowest is i1 and the highest is in.
  • Then, the median error to the reference value (m) is calculated using the equation:

  • m=i n/2
  • Where n is the number of properties used and i is the list of those errors listed in ascending order by absolute value.
  • Absolute Mean Error
  • The absolute mean error (a) is calculated using the equation:

  • a=Σ(|s n −r n |/s n)/n
  • Where
  • sn is the sale price or new data price for property n,
  • rn is the reference value price for property n, and
  • n is the number of properties used in the calculation.
  • The summation in the previous equation is calculated from 1 to n, where 1 is the first subject property and n is the last.
  • Mean Squared Error
  • The mean squared error (s) is calculated using the equation:

  • s=Σ[(s n −r n)2 /s n ]/n
  • Where sn is the sale price or new data price for property n, where rn is the reference value price for property n, and where n is the number of properties used in the calculation.
  • Automated Valuation Model Accuracy
  • Automated valuation model accuracy (c) is calculated using the equation:

  • c=n′/n
  • Where n′ is the number of properties whose valuation was within a predetermined percentage of the reference value, such as 10% of the reference value, and where n is the total number of properties. This percentage may be greater or smaller without affecting the method of this invention.
  • Percentage Incidence of Outlier Values
  • Percentage incidence of outlier values (o) is calculated using the following equation:

  • o=o′/n
  • Where o′ is the number of properties whose valuation was 25% more or less than the reference value and n is the total number of properties. The percentage that is considered to be an outlier value may be greater or smaller without affecting the method of this invention.
  • Each of the above calculations results in a decimal number that is representative of a percentage value. Depending on the indicator of accuracy a large percentage may indicate accuracy or may indicate inaccuracy. A large percentage, for example in hit rate percentage, indicates accuracy. In one embodiment, the highest hit rate percentage is ranked number one in a particular geographic area, economic tier, or property type. In other indicators of accuracy, for example absolute mean error, a lower percentage indicates more accuracy, stating that the average percentage of value that this particular automated valuation model “misses” the target. For this indicator and in this embodiment, the lowest percentage is ranked highest. In each of the categories, each automated valuation model is ordinally ranked as compared with every other automated valuation model. In this embodiment, these rankings are then multiplied, for each of the automated valuation models in each geographic area, economic tier, or property type, by the associated weighting factor. Each of these multiplied values are added together and multiplied by a scaling factor, such as ten, which results in an “accuracy score.” The lowest accuracy score in a particular geographic area, economic tier or property type is the most accurate automated valuation model.
  • Additionally, once the automated valuation models have been ranked, according the above process, a secondary ranking may take place. During this ranking, the same reference values, usually sale prices, but sometimes alternative new price data, are considered. The computation of rankings is performed using data for which the primary ranked AVM did not return an evaluation. For example, the hit rate is calculated for the primary ranking using all reference values and automated valuation model valuations for each subject property. The secondary hit rate is calculated using only reference values and automated valuation model valuations for subject properties for which the primary automated valuation model did not return a valuation. Similarly, the absolute mean error would only be calculated for properties for which there was no valuation returned by the primary ranked automated valuation model in that geographic area, economic tier, or property type. The purpose of this secondary ranking is to help enable automated valuation model users to determine which automated valuation model to use second, if the primary ranked automated valuation model is unable to value a particular subject property.
  • In the next step, any external client databases 136 are updated. This connection is made through the input and output connectors 110. The external client database 136 is a database of a client to which the new rankings and accuracy indicators and weights are sent. Additionally, the new data generated by the calculation processor 106 and the comparison processor 107 is sent through the input and output connectors 110 and appended to the AVM accuracy database 120. Alternatively, the rankings, reference values and automated valuations may be added to a new AVM accuracy database 120 for only a particular time period, such as a three-month period. In either database update, the database 120 is updated to reflect changes in the rankings of automated valuation models and the numerical changes to the various indicators of accuracy. Additional data users 138 may also be updated through the input and output connectors 110. These may include the users of the raw data, such as the indicators of accuracy or simply raw data from the new data and valuation database 116 to be used to perform their own calculations. Finally, the new data source 114 will likely be used by relevant public agencies to update public record databases 132. Additionally, various automated valuation model venders periodically use this new data, generally on a monthly or quarterly basis, to update the databases of each of the automated valuation models 134.
  • FIG. 3, depicts the steps involved in the data collection process of the present invention. The first step in this process is to gather or generate new property value indication data 140. This entails the reception of new sale data from publicly available records. In the preferred method of this invention, the data is appended to the database as quickly as it is received. Next, this new data is received and stored 142 in the new data and valuation database 116. Next, valuation requests are made to each automated valuation model to be tested 144. This request need not occur immediately after receipt of new sale data. In the preferred embodiment of the invention, the new sale data is aggregated and all automated valuations are requested at the end of a day. In alternative embodiments, the requests 144 are made immediately upon receipt of new sale data or at any other predetermined time interval such as bi-daily or weekly. Using either embodiment, the database which stores the automated valuation model valuations and the actual sale price or other reference data is continuously updated and becomes increasingly useful for deriving measures of the accuracy of each automated valuation model.
  • In the preferred embodiment, the valuation data provided by each automated valuation model is done without reference to the very recent sale. New sale data is gathered using this method for both the purpose of testing automated valuation models but also to provide that new sale data to automated valuation models. Most automated valuation model databases are only updated monthly or quarterly with the data provided using the aggregate new sale data gathered using this method. Therefore, each automated valuation model will most likely provide an automated valuation for the new sale data subject property without reference to the most recent sale price because it has not yet been provided. Even if a particular automated valuation model is updated on a weekly or even a daily basis, the present invention preferably provides accuracy testing of the automated valuation models prior to providing data of the most recent property sales for updates to automated valuation models. This enables the method of this invention to most effectively gauge the accuracy of each automated valuation model being tested. Once each request is made and each automated valuation model responds with its valuation, this data is received and stored 146 into the same database as the new data. Alternatively, this data may be stored in a separate database from that of the new data or in individual databases for each automated valuation model being tested.
  • Referring now to FIG. 4, a flowchart depicting the steps involved in the periodic ranking and AVM accuracy database update are depicted. The first step is to calculate the accuracy indicators 148. This takes place in the preferred embodiment in the calculation processor 106. Each of the indicators of accuracy are calculated for each automated valuation model. This is done using the aggregate reference values and automated valuations for each subject property over a set period of time. This time period may be over the life of a particular automated valuation model or for using all data that the method of this invention has been able to store. Alternatively, these indicators of accuracy may be calculated for only a smaller subset of all of the data available because of the method of this invention. Each of the indicators of accuracy is a percentage or an average percentage of a plurality of valuations that may be used to show, in general, the accuracy of each of the tested automated valuation models.
  • Next, in one embodiment, each AVM is ranked in each indicator of accuracy and in each category-geographic area, economic tier, or property type-by ordinal numbers, then the ordinal number for each AVM's ranking is multiplied by its corresponding weighting factor, and the products of these multiplications are summed 150. The corresponding sum of the products is then used to rank the automated valuation models 152 in the embodiments described. In the embodiment that uses ordinal rankings multiplied by weighting factors, the lowest summed number is the most accurate automated valuation model, and the next lowest summed number is the next most accurate.
  • In the alternative embodiment that uses a weighting equation, an overall accuracy indicating score is created using an additional formula, rather than ranking each automated valuation model compared to another. In this embodiment, each of the accuracy indicators is calculated in much the same way, but they are weighted using a formula or equation. In the alternative embodiment, the highest calculated number is the most accurate automated valuation model, and the next highest calculated number is the next most accurate. The automated valuation models are then ranked according to which has the highest accuracy score. This alternative embodiment will be described further in discussing FIGS. 7 a-8 b.
  • Finally, in the embodiments described, once the rankings are calculated, the method of this invention updates the AVM accuracy database 154. This database then contains the most up to-date rankings and the indicators that led to those rankings. Referring now to FIG. 5 a, the contents of a primary ranking database are depicted. The various indicators of accuracy are depicted in column 155. These are hit rate percentage 156, median error 158, absolute mean error 160, mean squared error 162, AVM accuracy 164, and outlier percentage 166. Hit rate percentage 156 refers to the percentage of reference properties for which each automated valuation model was able to return a valuation. Median error 158 refers to the percentage error at the middle range of the ordered “list” of errors with reference to the reference value. Absolute mean error 160 refers to the percentage average of the absolute value of the list of errors in valuation. Mean squared error 162 is the percentage average in relation to the reference value of the squares of the “list” of errors in valuation. AVM accuracy 164 refers to the percentage of automated valuations within plus or minus 10% of the reference value. Outlier percentage 166 refers to the percentage of automated valuations that are more than plus or minus 25% of the reference value.
  • Also depicted in element 168 in the embodiment shown in FIG. 5 a is the score or accuracy score of the particular automated valuation model in the particular geographic area. The rank for each evaluated AVM based on factors 156 through 166 for a predetermined geographic area is in column 170. The rankings for three AVMs, X, Y, and Z, are shown in FIG. 5 a for three states, Florida, California, and New Jersey. A lower ranking number indicates higher accuracy. For example, the rankings for three AVMs for Florida are shown, AVM X 172, AVM Y 174, and AVM Z 176. The various factors are compared and then ranked. For example in element 178, the hit rate percentage of AVM X in Florida is 93.6%. For AVM Y in Florida, depicted in element 180, the hit rate percentage is 85.1%. For AVM Z in Florida, the percentage is 67.8%. Because the higher the hit rate, that is the number of properties for which a response was generated by the automated valuation model, the higher the accuracy, a larger percentage for this factor is given a higher rank. Therefore, AVM X is ranked #1 in element 184, AVM Y is ranked #2 in element 186 and AVM Z is ranked #3 in element 188.
  • Similarly, AVM Accuracy, depicted in element 164, is returned as a percentage. This represents, as stated above, the percentage of values that were within plus or minus 10% of the actual price or other reference value. The percentage for AVM X of accuracy was 49.2%, depicted in element 190. The same percentage for AVM Y was 53.4%, depicted in element 192. The same percentage for AVM Z was 50.5%, depicted in element 194. This results in a ranking of #3 for AVM X in element 196, of #1 for AVM Y in element 198 and a ranking of #2 for AVM Z in element 200.
  • These rankings are then aggregated using a particular weighting applied to each factor. In alternative embodiments, as stated above, different factors may be used. Additionally, alternative weightings may be used in order to improve the usefulness of the rankings or to provide different types of results. Specific alternative embodiments may be implemented to provide a particular type of result. In the first embodiment, the weighting factors are as follows:
  • hit rate is multiplied by 0.05 or 5%
  • median error is multiplied by 0.1 or 10%
  • absolute mean error is multiplied by 0.1 or 10%
  • mean error squared is multiplied by 0.1 or 10%
  • automated valuation model accuracy is multiplied by 0.35 or 35%
  • outlier percentage is multiplied by 0.30 or 30%
  • The results of these multiplications are then added together and that sum is multiplied by a scaling factor such as ten to arrive at an accuracy score or score as depicted in element 168.
  • For example, the scaled result for AVM X in Florida, with the weighting factors multiplied by their corresponding rankings, is computed as follows:

  • Score=10*((0.05*1)+(0.10*3)+(0.10*3)+(0.10 3)+(0.35*3)+(0.30*1))

  • which equals

  • 10*(0.05+0.30+0.30+0.30+1.05+0.30)
  • which equals
  • 23 which is the reported score in element 202.
  • Using the same computation procedure, the score for AVM Y in Florida is 13.5, as depicted in element 204. Similarly, the score for AVM Z in Florida is 23.5, as depicted in element 206.
  • The same factors, weightings and scores are calculated for each geographic area, economic tier, or property type. Other states, such as California or New Jersey are depicted in elements 208 and 210 respectively. The rankings and calculated percentages for both states are also depicted in the remainder of this table. The contents of this table may be all or part of the content delivered to third party requesters. However, the database of properties valuations and reference values is maintained, along with the data depicted here in this embodiment. The embodiments described update the database quarterly or monthly. Many other automated valuation models may have data stored and ranked accordingly using the method of this invention.
  • Referring now to FIG. 5 b, the contents of an exemplary secondary ranking database are depicted. A secondary ranking provides a ranking of the automated valuation models for properties for which the automated valuation model ranked first in the primary ranking was unable to return a valuation. This ranking is therefore useful in determining which automated valuation model is a good “second choice.” This database represents most of the same calculations as those of the primary raking, minus the data that was already calculated using the primary ranked automated valuation model, that is the automated valuation model that ranked number 1 overall in a particular category or geographic area. Similar factors are considered along the left column. These are hit rate percentage 212, median error 214, absolute mean error 216, mean squared error 218, AVM accuracy 220 and outlier percentage 222. This secondary ranking is used to determine which automated valuation model to use after the primary ranked model. There is an importance in providing any valuation after the primary ranked automated valuation model was unable to provide one. The hit rate percentage of 70.3% is depicted, as shown in element 213. Also depicted is the score 224. Along the top, the rank 226 or 230, the automated valuation model being tested, such as AVM X—Florida 228 or AVM Z—Florida 232 are depicted. In the first embodiment, as indicated above, the weighting factors are as follows:
  • hit rate is multiplied by 0.05 or 5%
  • median error is multiplied by 0.1 or 10%
  • absolute mean error is multiplied by 0.1 or 10%
  • mean error squared is multiplied by 0.1 or 10%.
  • automated valuation model accuracy is multiplied by 0.35 or 35%
  • outlier percentage is multiplied by 0.30 or 30%
  • The results of these multiplications are then added together and that sum is multiplied by a scaling factor such as ten to arrive at an accuracy score or score as depicted in element 224.
  • For example, the scaled result for AVM X in Florida, with the weighting factors multiplied by their corresponding rankings, is computed as follows:

  • Score=10*((0.05*1)+(0.10*1)+(0.10.*1)+(0.10*1)+(0.35 1)+(0.30*1))

  • which equals

  • 10*(0.05+0.10+0.10+0.10+0.35+0.30)
  • which equals
  • 10 which is the reported score in element 234.
  • Using the same computation procedure, the secondary ranking score for AVM Z in Florida is 20, as depicted in element 236.
  • Referring now to FIG. 6 a, a summary of the example primary weighted ranking is depicted. This is a summary form of the results tabulated in FIG. 5 a. In the left column, each automated valuation model 238 is depicted. The Florida ranking 240 is shown at the top. The score 242, from FIG. 5 a and rankings, based upon the score from FIG. 5 a are depicted. The automated valuation model with the lowest score in a particular area is the most accurate in that area, according to one embodiment of this invention. For example, AVM Y 246 with a score of 13.5, depicted in element 248 is ranked #1 in element 250. AVM X 252 with a score of 23, depicted in element 254 is ranked #2 in element 256. AVM Z 258 with a score of 23.5, depicted in element 260 is ranked #3 in element 262. This consolidated table can be used to describe the relative accuracies of each automated valuation model in relation to each other in a particular geographic area (in this case) or other such division of properties. Many other automated valuation models may be included in this table and ranked accordingly.
  • Referring now to FIG. 6 b, an example summary secondary weighted ranking is depicted. This table is similar to the table depicted in FIG. 6 a, but is based on the data contained in FIG. 5 b. Therefore, this data excludes the valuations that were theoretically done using the primary ranked automated valuation model that is the most accurate one in a particular geographic area. This weighted ranking displays in the tabular form, automated valuation model 264, score 266, and rank 268. The Florida ranking is depicted in element 270. In Florida, AVM X 272 had a secondary score of 10, depicted in element 274 and a ranking of #1, depicted in element 276. All of the remaining automated valuation models are ranked in this secondary ranking.
  • Referring now to FIG. 7 a, the results of an alternative embodiment of the method used to implement the primary ranking of this invention is depicted. In this embodiment of the present invention, a formula is used to provide the relative weightings of the individual indicators of accuracy. The elements depicted in FIG. 7 a are much the same as those depicted in FIG. 5 a. The indicators of accuracy hit rate percentage 278 through outlier percentage 280 are the same. The values for hit rate percentage 278 in each automated valuation model are the same as in FIG. 5 a. For example, the hit rate percentage for AVM X in Florida is 93.6% as depicted in element 288. Similarly, the outlier percentage, depicted in element 290 is 12.7%. Each of the same automated valuation models are depicted; AVM X in 282, AVM Y in 284 and AVM Z in 286. The formulas used to calculate each of the percentage values are the same. However, an overall accuracy indicating score 292 is created using an additional formula, rather than ranking each automated valuation model's accuracy indicators compared to another. The following is the score computation equation where each percentage is represented as a decimal:

  • Score=f*((h+c)−(m+a+s+o))
  • Where
  • f=multiplier factor, in the preferred embodiment it is 500
  • h=hit rate percentage
  • c=automated valuation model accuracy
  • m=median error to the reference value
  • a=absolute mean error s=mean squared error
  • o=percentage incidence of outlier values
  • Using this formula, along with the data in element 288 through element 290, the score in element 294 is calculated. The equation, with the numbers in place of the variables would appear as follows:

  • Score=500*((0.936+0.492)−(0.073+0.054+0.003+0.127))

  • which equals

  • 500*(1.428−0.257)
  • which equals
  • 585 which is the reported score in element 294.
  • This equation is used to calculate each of the scores for the automated valuation models in each geographic area, economic tier or property type. A multiplier factor of 500 is used in the invention to accentuate the differences in accuracy between the automated valuation model valuations. A higher score represents a more accurate automated valuation in a particular geographic area, economic tier or property type. In Florida, for example AVM X is the most accurate with a score of 585, depicted in element 294. AVM Y is the next-most accurate with a score of 566, depicted in element 296 and AVM Z is the least accurate with a score of 455, depicted in element 298.
  • Referring now, to FIG. 7 b, the results of an alternative embodiment of the method used to implement the secondary ranking of this invention is depicted. As stated above, the secondary ranking is useful in finding a “second choice” automated valuation model. The valuations performed by the first-ranked automated valuation model in the primary ranking are removed. In this example, the “overall” primary automated valuation model was removed. This means that because AVM Y was the most accurate overall, it was “chosen” as the primary automated valuation model for each region. In an alternative method, the primary automated valuation model may be chosen for a particular geographic region, economic tier or property type. In that embodiment, for example, in Florida, AVM X would have been the primary automated valuation model because it was more accurate in that geographic area. So, it would be excluded from the secondary ranking.
  • In this embodiment, only the valuations for which the top ranked primary automated valuation model, AVM Y, did not return a value are considered. In this table many of the same elements depicted in FIG. 5 b are shown. Indicators of accuracy from hits 300 to outlier percentage 302 are depicted. Both of the previous non-top-ranked automated valuation models are included, AVM X 304 and AVM Z 306.
  • As stated above, hit rate percentage is again depicted because for a secondary ranking, the percentage of additional properties for which valuations are returned is important in deciding which automated valuation model to use as a “second choice.” In this table, the hit rate percentage for AVM X in Florida was 70.3%, depicted in element 308. In an alternative embodiment of this implementation, the total number of the remaining properties that were able to be valued may be considered.
  • The same equation used for the primary ranking is used again, excluding the previously valued properties. The equation is as follows:

  • Score=f*((h+c)−(m+a+s+o))
  • Where
  • f=multiplier factor, in the preferred embodiment it is 500
  • h=hit rate percentage
  • c=automated valuation model accuracy
  • m=median error to the reference value
  • a=absolute mean error
  • s=mean squared error
  • o=percentage incidence of outlier values
  • The equation for AVM X in Florida with the numbers in place of the variables would appear as follows:

  • Score=500*((0.703+0.571)−(0.098+0.077+0.006+0.238))

  • which equals

  • 500*(1.274−0.419)
  • which equals
  • 427 which is the reported score in element 316.
  • As stated above, a higher score denotes a higher level of accuracy. The score of AVM Z in Florida, depicted in element 318 is only −373 and it is therefore considerably less accurate than AVM X in Florida in the properties remaining after the top ranked automated valuation model in the primary ranking has valued properties. Both the primary and secondary ranking of automated valuation model accuracy using either of the embodiments provides a means by which automated valuation models may be compared against each other for accuracy that are improvements over the prior art.
  • Referring now to FIG. 8 a, an example summary primary weighted ranking of the alternative embodiment is depicted. This is a summary form of the data presented in FIG. 7 a depicting only the scores in each location and the relative ranking. Similar to FIG. 6 a above, the state for which the rankings are being done is depicted as Florida in element 320. The column for automated valuation model 322, includes AVM X 328, AVM Y 330 and AVM Z 332. The respective scores for each of these automated valuation models are 585, depicted in element 334; 566, depicted in element 336; and 455, depicted in element 338. Therefore, AVM X 328 with its score of 585 in Florida is ranked number one, as depicted in element 340. The overall rankings depicted below are a result of the addition of each of the scores in each area. AVM Y 342 with its overall score of 1861, depicted in element 344 is ranked number one 346 overall. In an alternative embodiment, the overall ranking may be determined by dividing the geographic area score by the number of properties in that area for which automated valuations were provided and then adding the numbers together. This would provide a ranking weighted as to the percentage of homes in a particular market or its relative importance in use of automated valuation data.
  • Referring now to FIG. 8 b, a secondary weighted ranking is depicted. This is a summary table of the data contained in FIG. 7 b. This table is useful in comparing directly secondary rankings. As in FIG. 6 b, the automated valuation model being ranked is depicted in the column denoted by element 350. The Florida secondary ranking, for example, is depicted in element 348. AVM X in Florida, for example, depicted in element 354, received a rank of one, as depicted in element 356. The column depicted in element 352 denotes the rank. Overall, AVM X 358, received the number one ranking 360 with a score of 1079, depicted in element 362.
  • A method of and apparatus for real time testing of automated valuation models has been described. It is to be understood that the foregoing description has been made with respect to specific embodiments thereof for illustrative purposes only. The overall spirit and scope of the present invention is limited only by the following claims, as defined in the foregoing description.

Claims (21)

1. (canceled)
2. A method comprising:
providing ranking criteria to an automated valuation model testing entity;
accessing, by a computer system comprising computer hardware, rankings data and accuracy data associated with a plurality of automated valuation models, and provided by the automated valuation model testing entity, wherein the rankings data comprises rankings of the plurality of automated valuation models based at least in part on the provided ranking criteria and comparisons of valuations for a subject property received from the plurality of automated valuation models without reference to new valuation data associated with the subject property; and
performing, by the computer system, one or more calculations based at least in part on the accessed rankings data.
3. The method of claim 2, wherein the ranking criteria comprises cost preferences associated with automated valuation models.
4. The method of claim 2, wherein said accuracy data comprises data indicative of accuracy associated with each automated valuation model.
5. The method of claim 2, wherein the accuracy data comprises weightings associated with the accuracy data.
6. The method of claim 2, wherein the automated valuation model testing entity provides the rankings data in real-time.
7. The method of claim 2, wherein the automated valuation model testing entity provides the rankings data periodically.
8. The method of claim 2, wherein the new valuation data comprises property sale data.
9. The method of claim 2, wherein the new valuation data comprises an appraisal for the subject property.
10. The method of claim 2, wherein the new valuation data comprises an automated valuation of the subject property.
11. A system comprising:
physical data storage configured to store (1) a reference valuation for a subject property and (2) a plurality of automated valuations of the subject property, each automated valuation generated by an automated valuation model (AVM) without use of the reference valuation for the subject property, wherein different automated valuations correspond to different AVMS; and
a computer system in communication with the physical data storage, the computer system comprising computer hardware, the computer system programed to:
provide ranking criteria to an automated valuation model testing entity;
access rankings data and accuracy data associated with a portion of the plurality of automated valuation models, and provided by the automated valuation model testing entity, wherein the rankings data comprises rankings of the portion of the plurality of automated valuation models based at least in part on the provided ranking criteria and comparisons of valuations for the subject property received from the portion of the plurality of automated valuation models; and
perform one or more calculations based at least in part on the accessed rankings data.
12. The system of claim 11, wherein the ranking criteria comprises cost preferences associated with automated valuation models.
13. The system of claim 11, wherein the automated valuation model testing entity provides the rankings data in real-time.
14. The system of claim 11, wherein the automated valuation model testing entity provides the rankings data periodically.
15. The system of claim 11, wherein the reference valuation comprises property sale data.
16. The system of claim 11, wherein the reference valuation comprises an appraisal for the subject property.
17. The system of claim 11, wherein the reference valuation comprises an automated valuation of the subject property.
18. A tangible computer-readable medium that stores thereon a plurality of computer-executable instructions configured, when executed by a computer processor, to cause computer hardware to perform operations comprising:
accessing rankings data and accuracy data associated with a plurality of automated valuation models, and provided by the automated valuation model testing entity, wherein the rankings data comprises rankings of the plurality of automated valuation models based at least in part on the provided ranking criteria and comparisons of valuations for a subject property received from the plurality of automated valuation models without reference to new valuation data associated with the subject property;
testing the rankings data to determine accuracy of the rankings data; and
performing one or more calculations based at least in part on the accessed rankings data.
19. The tangible computer-readable medium of claim 18, wherein the automated valuation model testing entity provides the rankings data in real-time.
20. The tangible computer-readable medium of claim 18, wherein the new valuation data comprises property sale data.
21. The tangible computer-readable medium of claim 18, wherein the reference valuation comprises an appraisal for the subject property.
US13/756,927 2004-12-08 2013-02-01 Method and apparatus for testing automated valuation models Abandoned US20130246287A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/756,927 US20130246287A1 (en) 2004-12-08 2013-02-01 Method and apparatus for testing automated valuation models

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US11/007,750 US8001024B2 (en) 2004-12-08 2004-12-08 Method and apparatus for testing automated valuation models
US13/182,289 US8370239B2 (en) 2004-12-08 2011-07-13 Method and apparatus for testing automated valuation models
US13/756,927 US20130246287A1 (en) 2004-12-08 2013-02-01 Method and apparatus for testing automated valuation models

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/182,289 Continuation US8370239B2 (en) 2004-12-08 2011-07-13 Method and apparatus for testing automated valuation models

Publications (1)

Publication Number Publication Date
US20130246287A1 true US20130246287A1 (en) 2013-09-19

Family

ID=36575548

Family Applications (3)

Application Number Title Priority Date Filing Date
US11/007,750 Active 2030-04-17 US8001024B2 (en) 2004-12-08 2004-12-08 Method and apparatus for testing automated valuation models
US13/182,289 Active US8370239B2 (en) 2004-12-08 2011-07-13 Method and apparatus for testing automated valuation models
US13/756,927 Abandoned US20130246287A1 (en) 2004-12-08 2013-02-01 Method and apparatus for testing automated valuation models

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US11/007,750 Active 2030-04-17 US8001024B2 (en) 2004-12-08 2004-12-08 Method and apparatus for testing automated valuation models
US13/182,289 Active US8370239B2 (en) 2004-12-08 2011-07-13 Method and apparatus for testing automated valuation models

Country Status (5)

Country Link
US (3) US8001024B2 (en)
AU (1) AU2005314213A1 (en)
GB (1) GB2437860A (en)
NZ (1) NZ555315A (en)
WO (1) WO2006062949A2 (en)

Families Citing this family (80)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6615189B1 (en) 1998-06-22 2003-09-02 Bank One, Delaware, National Association Debit purchasing of stored value card for use by and/or delivery to others
US7809642B1 (en) 1998-06-22 2010-10-05 Jpmorgan Chase Bank, N.A. Debit purchasing of stored value card for use by and/or delivery to others
US6032136A (en) 1998-11-17 2000-02-29 First Usa Bank, N.A. Customer activated multi-value (CAM) card
US7660763B1 (en) 1998-11-17 2010-02-09 Jpmorgan Chase Bank, N.A. Customer activated multi-value (CAM) card
US8793160B2 (en) 1999-12-07 2014-07-29 Steve Sorem System and method for processing transactions
US7313546B2 (en) 2001-05-23 2007-12-25 Jp Morgan Chase Bank, N.A. System and method for currency selectable stored value instrument
AU2002327322A1 (en) 2001-07-24 2003-02-17 First Usa Bank, N.A. Multiple account card and transaction routing
US7311244B1 (en) * 2001-08-13 2007-12-25 Jpmorgan Chase Bank, N.A. System and method for funding a collective account by use of an electronic tag
US8020754B2 (en) 2001-08-13 2011-09-20 Jpmorgan Chase Bank, N.A. System and method for funding a collective account by use of an electronic tag
US7899753B1 (en) 2002-03-25 2011-03-01 Jpmorgan Chase Bank, N.A Systems and methods for time variable financial authentication
US8751391B2 (en) 2002-03-29 2014-06-10 Jpmorgan Chase Bank, N.A. System and process for performing purchase transactions using tokens
US20040210498A1 (en) 2002-03-29 2004-10-21 Bank One, National Association Method and system for performing purchase and other transactions using tokens with multiple chips
US8239304B1 (en) 2002-07-29 2012-08-07 Jpmorgan Chase Bank, N.A. Method and system for providing pre-approved targeted products
US7809595B2 (en) 2002-09-17 2010-10-05 Jpmorgan Chase Bank, Na System and method for managing risks associated with outside service providers
US8306907B2 (en) 2003-05-30 2012-11-06 Jpmorgan Chase Bank N.A. System and method for offering risk-based interest rates in a credit instrument
US7953663B1 (en) 2003-09-04 2011-05-31 Jpmorgan Chase Bank, N.A. System and method for financial instrument pre-qualification and offering
US7599882B2 (en) * 2003-11-14 2009-10-06 First American Corelogic, Inc. Method for mortgage fraud detection
US20050171822A1 (en) * 2004-02-03 2005-08-04 First American Real Estate Solutions, L.P. Responsive confidence scoring method for a proposed valuation of aproperty
US7392222B1 (en) 2004-08-03 2008-06-24 Jpmorgan Chase Bank, N.A. System and method for providing promotional pricing
US20060085234A1 (en) * 2004-09-17 2006-04-20 First American Real Estate Solutions, L.P. Method and apparatus for constructing a forecast standard deviation for automated valuation modeling
US8001024B2 (en) 2004-12-08 2011-08-16 Corelogic Information Solutions, Inc. Method and apparatus for testing automated valuation models
US8386395B1 (en) * 2005-04-28 2013-02-26 Federal Home Loan Mortgage Corporation (Freddie Mac) Systems and methods for modifying a loan
US7853518B2 (en) * 2005-05-24 2010-12-14 Corelogic Information Solutions, Inc. Method and apparatus for advanced mortgage diagnostic analytics
US7401731B1 (en) 2005-05-27 2008-07-22 Jpmorgan Chase Bank, Na Method and system for implementing a card product with multiple customized relationships
US20070033122A1 (en) * 2005-08-04 2007-02-08 First American Real Estate Solutions, Lp Method and apparatus for computing selection criteria for an automated valuation model
WO2007019326A2 (en) 2005-08-05 2007-02-15 First American Corelogic Holdings, Inc. Method and system for updating a loan portfolio with information on secondary liens
US8676680B2 (en) 2006-02-03 2014-03-18 Zillow, Inc. Automatically determining a current value for a home
US8515839B2 (en) 2006-02-03 2013-08-20 Zillow, Inc. Automatically determining a current value for a real estate property, such as a home, that is tailored to input from a human user, such as its owner
US8408455B1 (en) 2006-02-08 2013-04-02 Jpmorgan Chase Bank, N.A. System and method for granting promotional rewards to both customers and non-customers
US7784682B2 (en) 2006-02-08 2010-08-31 Jpmorgan Chase Bank, N.A. System and method for granting promotional rewards to both customers and non-customers
US7753259B1 (en) 2006-04-13 2010-07-13 Jpmorgan Chase Bank, N.A. System and method for granting promotional rewards to both customers and non-customers
US20080077458A1 (en) 2006-09-19 2008-03-27 Andersen Timothy J Collecting and representing home attributes
US20080255862A1 (en) * 2007-04-11 2008-10-16 Bailey Gregory A Predictive asset ranking score of property
US8676642B1 (en) 2007-07-05 2014-03-18 Jpmorgan Chase Bank, N.A. System and method for granting promotional rewards to financial account holders
US8417601B1 (en) 2007-10-18 2013-04-09 Jpmorgan Chase Bank, N.A. Variable rate payment card
US11244289B2 (en) * 2007-11-02 2022-02-08 Citicorp Credit Services, Inc. (Usa) Methods and systems for managing financial institution customer accounts
US8140421B1 (en) 2008-01-09 2012-03-20 Zillow, Inc. Automatically determining a current value for a home
US20100274708A1 (en) * 2008-05-29 2010-10-28 Allen Lewis J Apparatus and method for creating a collateral risk score and value tolerance for loan applications
US20110119161A1 (en) * 2009-11-18 2011-05-19 Van Treeck George M Automated ratings of new products and services
US10380653B1 (en) 2010-09-16 2019-08-13 Trulia, Llc Valuation system
US10460406B1 (en) 2011-03-09 2019-10-29 Zillow, Inc. Automatically determining market rental rates for properties
US10198735B1 (en) 2011-03-09 2019-02-05 Zillow, Inc. Automatically determining market rental rate index for properties
US20120303536A1 (en) * 2011-05-25 2012-11-29 Corelogic Information Solutions, Inc. Property complexity scoring system, method, and computer program storage device
US20130144796A1 (en) * 2011-12-06 2013-06-06 Fannie Mae Assigning confidence values to automated property valuations by using the non-typical property characteristics of the properties
US10672088B2 (en) * 2012-06-19 2020-06-02 Fannie Mae Automated valuation model with comparative value history information
US9310204B2 (en) * 2013-02-04 2016-04-12 The Boeing Company Route modeler for improving desired environmental and economic flight characteristics
US20140257924A1 (en) * 2013-03-08 2014-09-11 Corelogic Solutions, Llc Automated rental amount modeling and prediction
US10754884B1 (en) 2013-11-12 2020-08-25 Zillow, Inc. Flexible real estate search
US10984489B1 (en) 2014-02-13 2021-04-20 Zillow, Inc. Estimating the value of a property in a manner sensitive to nearby value-affecting geographic features
US9959560B1 (en) 2014-08-26 2018-05-01 Intuit Inc. System and method for customizing a user experience based on automatically weighted criteria
AU2015309650A1 (en) * 2014-08-26 2017-03-09 Municipal Property Assessment Corporation Method and system for real estate valuation
US11354755B2 (en) 2014-09-11 2022-06-07 Intuit Inc. Methods systems and articles of manufacture for using a predictive model to determine tax topics which are relevant to a taxpayer in preparing an electronic tax return
US11093982B1 (en) 2014-10-02 2021-08-17 Zillow, Inc. Determine regional rate of return on home improvements
US10096072B1 (en) 2014-10-31 2018-10-09 Intuit Inc. Method and system for reducing the presentation of less-relevant questions to users in an electronic tax return preparation interview process
US10013721B1 (en) 2014-10-31 2018-07-03 Intuit Inc. Identification of electronic tax return errors based on declarative constraints
US10255641B1 (en) 2014-10-31 2019-04-09 Intuit Inc. Predictive model based identification of potential errors in electronic tax return
US20160171564A1 (en) * 2014-12-11 2016-06-16 Fannie Mae Subject appraisal discrepancy analysis
US20160180470A1 (en) * 2014-12-23 2016-06-23 Intuit Inc. Method and system for evaluating interchangeable analytics modules used to provide customized tax return preparation interviews
US10628894B1 (en) 2015-01-28 2020-04-21 Intuit Inc. Method and system for providing personalized responses to questions received from a user of an electronic tax return preparation system
US10643232B1 (en) 2015-03-18 2020-05-05 Zillow, Inc. Allocating electronic advertising opportunities
US10176534B1 (en) 2015-04-20 2019-01-08 Intuit Inc. Method and system for providing an analytics model architecture to reduce abandonment of tax return preparation sessions by potential customers
US10740853B1 (en) 2015-04-28 2020-08-11 Intuit Inc. Systems for allocating resources based on electronic tax return preparation program user characteristics
US10235630B1 (en) 2015-07-29 2019-03-19 Wells Fargo Bank, N.A. Model ranking index
US10740854B1 (en) 2015-10-28 2020-08-11 Intuit Inc. Web browsing and machine learning systems for acquiring tax data during electronic tax return preparation
US10937109B1 (en) 2016-01-08 2021-03-02 Intuit Inc. Method and technique to calculate and provide confidence score for predicted tax due/refund
US10475131B1 (en) 2016-01-27 2019-11-12 Intuit Inc. Methods, systems and computer program products for calculating an estimated result of a tax return
US10789549B1 (en) 2016-02-25 2020-09-29 Zillow, Inc. Enforcing, with respect to changes in one or more distinguished independent variable values, monotonicity in the predictions produced by a statistical model
US10410295B1 (en) 2016-05-25 2019-09-10 Intuit Inc. Methods, systems and computer program products for obtaining tax data
US10789278B1 (en) * 2016-06-30 2020-09-29 Costar Realty Information, Inc. Database search engine optimization
US10726509B1 (en) 2016-08-19 2020-07-28 Corelogic Solutions, Llc Data mining data records to determine networks and boundaries
US20190266681A1 (en) * 2018-02-28 2019-08-29 Fannie Mae Data processing system for generating and depicting characteristic information in updatable sub-markets
US10952947B2 (en) * 2018-03-27 2021-03-23 Elc Management Llc Thickened catalyzed dye system
US11369554B2 (en) * 2018-03-27 2022-06-28 Elc Management Llc Catalyzed dye system
US11861635B1 (en) 2019-03-20 2024-01-02 MFTB Holdco, Inc. Automatic analysis of regional housing markets based on the appreciation or depreciation of individual homes
US20200402116A1 (en) * 2019-06-19 2020-12-24 Reali Inc. System, method, computer program product or platform for efficient real estate value estimation and/or optimization
US11861748B1 (en) 2019-06-28 2024-01-02 MFTB Holdco, Inc. Valuation of homes using geographic regions of varying granularity
US11775897B2 (en) * 2019-07-02 2023-10-03 Adp, Inc. Predictive modeling method and system for dynamically quantifying employee growth opportunity
US20220172255A1 (en) * 2020-12-01 2022-06-02 Zillow, Inc. Confident processing of valuations from distributed models systems and methods
US11094135B1 (en) 2021-03-05 2021-08-17 Flyreel, Inc. Automated measurement of interior spaces through guided modeling of dimensions
US20220343377A1 (en) * 2021-04-23 2022-10-27 Wrench Ip Holding Company Automated pricing mechanism

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7836057B1 (en) * 2001-09-24 2010-11-16 Auguri Corporation Weighted preference inference system and method

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6609118B1 (en) 1999-06-21 2003-08-19 General Electric Company Methods and systems for automated property valuation
US7162445B2 (en) * 1999-12-30 2007-01-09 Ge Corporate Financial Services, Inc. Methods and systems for quantifying cash flow recovery and risk
US20020007336A1 (en) * 2000-04-04 2002-01-17 Robbins Michael L. Process for automated owner-occupied residental real estate valuation
US20010039506A1 (en) 2000-04-04 2001-11-08 Robbins Michael L. Process for automated real estate valuation
US6873972B1 (en) * 2000-08-01 2005-03-29 General Electric Company Systems and methods for credit line monitoring
US20020127529A1 (en) * 2000-12-06 2002-09-12 Cassuto Nadav Yehudah Prediction model creation, evaluation, and training
EP1410325A2 (en) * 2001-06-08 2004-04-21 Netuitive, Inc. Automated analyzers for estimation systems
US20030212565A1 (en) * 2002-05-08 2003-11-13 Badali Robert C. Appraisal and home inspection process
US7756804B2 (en) 2002-05-10 2010-07-13 Oracle International Corporation Automated model building and evaluation for data mining system
US20050171822A1 (en) * 2004-02-03 2005-08-04 First American Real Estate Solutions, L.P. Responsive confidence scoring method for a proposed valuation of aproperty
US20060015357A1 (en) * 2004-07-16 2006-01-19 First American Real Estate Solutions, L.P. Method and apparatus for spatiotemporal valuation of real estate
US20060085234A1 (en) 2004-09-17 2006-04-20 First American Real Estate Solutions, L.P. Method and apparatus for constructing a forecast standard deviation for automated valuation modeling
US8001024B2 (en) 2004-12-08 2011-08-16 Corelogic Information Solutions, Inc. Method and apparatus for testing automated valuation models

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7836057B1 (en) * 2001-09-24 2010-11-16 Auguri Corporation Weighted preference inference system and method

Also Published As

Publication number Publication date
US20060122918A1 (en) 2006-06-08
GB0713280D0 (en) 2007-08-15
WO2006062949A3 (en) 2007-05-10
US8001024B2 (en) 2011-08-16
US20120011075A1 (en) 2012-01-12
AU2005314213A1 (en) 2006-06-15
NZ555315A (en) 2009-12-24
US8370239B2 (en) 2013-02-05
GB2437860A (en) 2007-11-07
WO2006062949A2 (en) 2006-06-15

Similar Documents

Publication Publication Date Title
US8370239B2 (en) Method and apparatus for testing automated valuation models
US11449958B1 (en) Automatically determining a current value for a home
US8515839B2 (en) Automatically determining a current value for a real estate property, such as a home, that is tailored to input from a human user, such as its owner
US20190005553A1 (en) Automatically determining a current value for a home
US7970674B2 (en) Automatically determining a current value for a real estate property, such as a home, that is tailored to input from a human user, such as its owner
Nganje et al. Regional economic impacts of Fusarium head blight in wheat and barley
US8171022B2 (en) Methods, systems, and computer program products for facilitating user interaction with customer relationship management, auction, and search engine software using conjoint analysis
US20010039506A1 (en) Process for automated real estate valuation
US6792399B1 (en) Combination forecasting using clusterization
US20100023379A1 (en) Method and system for determining real estate market value changes
US20150012335A1 (en) Automated rental amount modeling and prediction
US20150193797A1 (en) Dynamic property pricing system for the real estate field and a method of performing the same
US20070192130A1 (en) System and method for rating service providers
US20110270625A1 (en) System, method and computer program for estimating medical costs
CN103412918A (en) Quality of service (QoS) and reputation based method for evaluating service trust levels
US20070033122A1 (en) Method and apparatus for computing selection criteria for an automated valuation model
WO2020029385A1 (en) Asset allocation rationality determining method and system, computer device, and storage medium
EP2646966A1 (en) Private company valuation
US20040210485A1 (en) Quoting strategy analysis system and method
Bosse et al. Transformation of forecasts for evaluating predictive performance in an epidemiological context
CN115391666A (en) Hospital online recommendation method and device, computer equipment and storage medium
JP6056061B2 (en) Information posting system
AU784943B2 (en) Loan processing system and method
JP5025384B2 (en) Energy transaction price forecasting system, forecasting method and program
US20230260035A1 (en) System and method for analyzing, evaluating and ranking properties using artificial intelligence

Legal Events

Date Code Title Description
AS Assignment

Owner name: BANK OF AMERICA, N.A., TEXAS

Free format text: SECURITY INTEREST;ASSIGNOR:CORELOGIC SOLUTIONS, LLC;REEL/FRAME:032798/0047

Effective date: 20140404

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: CORELOGIC SOLUTIONS, LLC, CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST RECORDED AT 032798/0047;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:056493/0957

Effective date: 20210604