SYSTEMS AND METHODS FOR PREDICTIVE QUERY IMPLEMENTATION AND USAGE IN A MULTI-TENANT DATABASE SYSTEM
Disclosed herein are systems and methods for predictive query implementation and usage in a multi-tenant database system including means for implementing predictive population of null values with confidence scoring, means for predictive scoring and reporting of business opportunities with probability to close scoring, and other related embodiments.
Latest Salesforce.com Patents:
This application is related to, and claims priority to, the provisional utility application entitled “SYSTEMS AND METHODS FOR PREDICTIVE QUERY IMPLEMENTATION AND USAGE IN A MULTI-TENANT DATABASE SYSTEM,” filed on Mar. 13, 2013, having an application number of 61/780,503 and attorney docket No. 8956P119Z (520PROV), the entire contents of which are incorporated herein by reference.
COPYRIGHT NOTICEA portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.
TECHNICAL FIELDEmbodiments of the invention relate generally to the field of computing, and more particularly, to systems and methods for predictive query implementation and usage in a multi-tenant database system including means for implementing predictive population of null values with confidence scoring, means for predictive scoring and reporting of business opportunities with probability to close scoring, and other related embodiments.
Embodiments are illustrated by way of example, and not by way of limitation, and will be more fully understood with reference to the following detailed description when considered in connection with the figures in which:
The subject matter discussed in the background section should not be assumed to be prior art merely as a result of its mention in the background section. Similarly, a problem mentioned in the background section or associated with the subject matter of the background section should not be assumed to have been previously recognized in the prior art. The subject matter in the background section merely represents different approaches, which in and of themselves may also correspond to embodiments of the claimed inventions.
Client organizations with datasets in their databases can benefit from predictive analysis. Unfortunately, there is no low cost and scalable solution in the marketplace today. Instead, client organizations must hire technical experts to develop customized mathematical constructs and predictive models which are very expensive. Consequently, client organizations without vast financial means are simply priced out of the market and thus do not have access to predictive analysis capabilities for their datasets.
The present state of the art may therefore benefit from methods, systems, and apparatuses for predictive query implementation and usage in a multi-tenant database system as described herein.
DETAILED DESCRIPTIONUsers wanting to perform predictive analytics and data mining against their datasets must normally hire technical experts and explain the problem they wish to solve and then turn their data over to the hired experts to apply specialized mathematical constructs in an attempt to solve the problem.
By analogy, many years ago when you designed a computer system it was necessary to also figure out how to put data on a physical disk. Now programmers do not concern themselves with such issues. Similarly, it is highly desirable to utilize a server and sophisticated database technology to perform data analytics for ordinary users without having to hire specialized experts. By doing so, resources could be freed up to focus on other problems.
Some machine learning capabilities exist today. For instance, present capabilities can answer questions such as, “Is this person going to buy product x?” But such simplistic technology is not sufficient for helping people to solve more complex problems. For instance, Kaiser Healthcare corporation with vast financial resources may be able to hire experts from KXEN to develop customized analytics to solve a Kaiser specific problem based on Kaiser's database, but a small company by contrast simply cannot afford to utilize KXEN's services as the cost far outweighs a small company's financial resources to do so. Thus, our exemplary small company would be forced to simply forgo solving the problem at hand.
Consider KXEN's own value proposition from their home page which states: “As a business analyst, you don't want to worry about complicated math or which algorithm to use. You need a model that is going to predict possible business outcomes. Is this customer likely to churn? Will they respond to a cross-sell or up-sell offer? . . . . We'll help you quickly get to the right algorithm for your business problem with a model built for accuracy and optimal results.”
If a small company lacks the financial resources to hire a company such as KXEN and lacks the technical know how to develop the “complicated math or [select] which algorithm to use,” then such company must go without.
Further still, the services offered today by technical experts in the field of analytics and predictive modeling provide solutions that are customized to the particular dataset of the customer. They do not offer capabilities that may be used by non-experts in an agnostic manner that is not anchored to a particular underlying dataset.
Veritable offers a predictive database and additional commands and verbs so that a non-expert user can query the predictive database with inquiries such as: “predict revenue from users where age is greater than 35.”
Further still, companies that hire analytics and predictive modeling are given a solution at adheres to the present data in their database, but do not adapt to changes in the data or the database structure over time. Thus, a large company may hire KXEN and math experts come in and stare at the data and build models, and so forth, and the models do work, but when the nature of the data changes over time within the database and layout of the data or the database's structure changes over time, as is normal and common for businesses, then the models stop working as they were customized for the particular data and database structure at a given point in time.
Because Veritable provides a predictive database that is not anchored to any particular underlying dataset, it remains useful as data and data structures change over time. For instance, data analysis performed by the Veritable core may simply be re-applied to a changed dataset. There is no need to re-hire experts or re-tool the models.
With respect salesforce.com specifically, the company offers cloud services to clients, organizations, and end users, and behind those cloud services is a multi-tenant database system which permits users to have customized data, customized field types, and so forth. The underlying data and data structures are customized by the client organizing for their own particular needs. Veritable may nevertheless be utilized on these varying datasets and data structures because it is not anchored to a particular underlying database scheme, structure, or content.
Customer organizations further benefit from the low cost of access. For instance, the cloud service provider may elect to provide the capability as part of an overall service offering at no additional cost, or may elect to provide the additional capabilities for an additional service fee. Regardless, because the Veritable capabilities are systematically integrated into the cloud service's computing architecture and do not require experts to custom tailor a solution to each particular client organizations' dataset and structure, the scalability brings massive cost savings, thus enabling our exemplary small company with limited financial resources to go from a 0% capability because they cannot afford to hire technical experts from KXEN to, for instance, a 95% accuracy capability using Veritable. Even if a large company with sufficient financial resources could feasibly hire KXEN to develop customized mathematical constructs and models, they would need to evaluate the ROI of hiring KXEN which may be able to get, for instance, 97% accuracy through customization but at high cost, versus using the turn-key access to the low cost cloud computing service which yields the exemplary 95% accuracy.
Regardless of the decision for a large company with financial means, a small company which would otherwise not have access to predictive analytic capabilities can benefit greatly as their capability for predictive analysis accuracy goes from 0% (e.g., mere guessing) to the exemplary 95% using the scalable architecture provided by Veritable.
In the following description, numerous specific details are set forth such as examples of specific systems, languages, components, etc., in order to provide a thorough understanding of the various embodiments. It will be apparent, however, to one skilled in the art that these specific details need not be employed to practice the embodiments disclosed herein. In other instances, well known materials or methods have not been described in detail in order to avoid unnecessarily obscuring the disclosed embodiments.
In addition to various hardware components depicted in the figures and described herein, embodiments further include various operations which are described below. The operations described in accordance with such embodiments may be performed by hardware components or may be embodied in machine-executable instructions, which may be used to cause a general-purpose or special-purpose processor programmed with the instructions to perform the operations. Alternatively, the operations may be performed by a combination of hardware and software.
Embodiments also relate to an apparatus for performing the operations disclosed herein. This apparatus may be specially constructed for the required purposes, or it may be a general purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions, each coupled to a computer system bus.
The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose systems may be used with programs in accordance with the teachings herein, or it may prove convenient to construct more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear as set forth in the description below. In addition, embodiments are not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the embodiments as described herein.
Embodiments may be provided as a computer program product, or software, that may include a machine-readable medium having stored thereon instructions, which may be used to program a computer system (or other electronic devices) to perform a process according to the disclosed embodiments. A machine-readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer). For example, a machine-readable (e.g., computer-readable) medium includes a machine (e.g., a computer) readable storage medium (e.g., read only memory (“ROM”), random access memory (“RAM”), magnetic disk storage media, optical storage media, flash memory devices, etc.), a machine (e.g., computer) readable transmission medium (electrical, optical, acoustical), etc.
Any of the disclosed embodiments may be used alone or together with one another in any combination. Although various embodiments may have been partially motivated by deficiencies with conventional techniques and approaches, some of which are described or alluded to within the specification, the embodiments need not necessarily address or solve any of these deficiencies, but rather, may address only some of the deficiencies, address none of the deficiencies, or be directed toward different deficiencies and problems where are not directly discussed.
In one embodiment, means for predictive query implementation and usage in a multi-tenant database system execute at an application in a computing device, a computing system, or a computing architecture, in which the application is enabled to communicate with a remote computing device over a public Internet, such as remote clients, thus establishing a cloud based computing service in which the clients utilize the functionality of the remote application which implements the predictive query and usage capabilities.
Model-based clustering techniques, including inference in Dirichlet process mixture models, have difficulty when different dimensions are best explained by very different clusterings. Based on MCMC inference in a novel nonparametric Bayesian model, methods automatically discover the number of independent nonparametric Bayesian models needed to explain the data, using a separate Dirichlet process mixture model for each group in an inferred partition of the dimensions. Unlike a DP mixture, the disclosed model is exchangeable over both the rows of a heterogeneous data array (the samples) and the columns (new dimensions), and can model any dataset as the number of samples and dimensions both go to infinity. Efficiency and robustness is improved through use of algorithms described herein which in certain instances require no preprocessing to identify veridical causal structure in provided raw datasets.
Clustering techniques are widely used in data analysis for problems of segmentation in industry, exploratory analysis in science, and as a preprocessing step to improve performance of further processing in distributed computing and in data compression. However, as datasets grow larger and noisier, the assumption that a single clustering or distribution over clusterings can account for all the variability in the observations becomes less realistic if not wholly infeasible.
From a machine learning perspective, this is an unsupervised version of the feature selection problem: different subsets of measurements should, in general, induce different natural clusterings of the data. From a cognitive science and artificial intelligence perspective, this issue is reflected in work that seeks multiple representations of data instead of a single monolithic representation.
As a limiting case, a robust clustering method should be able to ignore an infinite number of uniformly random or perfectly deterministic measurements. The assumption that a single nonparametric model must explain all the dimensions is partly responsible for the accuracy issues Dirichlet process mixtures often encounter in high dimensional settings. DP mixture based classifiers via class conditional density estimation highlight the problem. For instance, while a discriminative classifier can assign low weight to noisy or deterministic and therefore irrelevant dimensions, a generative model must explain them. If there are enough irrelevancies, it ignores the dimensions relevant to classification in the process. Combined with slow MCMC convergence, these difficulties have inhibited the use of nonparametric Bayesian methods in many applications.
To overcome these limitations, cross-categorization is utilized, which is an unsupervised learning technique for clustering based on MCMC inference in a novel nested nonparametric Bayesian model. This model can be viewed as a Dirichlet process mixture, over the dimensions or columns, of Dirichlet process mixture models over sampled data points or rows. Conditioned on a partition of the dimensions, our model reduces to an independent product of DP mixtures, but the partition of the dimensions, and therefore the number and domain of independent nonparametric Bayesian models, is also inferred from the data.
Standard feature selection results in the case where the partition of dimensions has only two groups. The described model utilizes MCMC because both model selection and deterministic approximations seem intractable due to the combinatorial explosion of latent variables, with changing numbers of latent variables as the partition of the dimensions changes.
The hypothesis space captured by the described model is super-exponentially larger than that of a Dirichlet process mixture, with a very different structure than a Hierarchical Dirichlet Process. A generative process, viewed as a model for heterogeneous data arrays with N rows, D columns of fixed type and values missing at random, can be described as follows:
1. For each dimension dεD:
-
- (a) Generate hyperparameters {right arrow over (λ)}d from an appropriate hyper-prior.
- (b) Generate the model assignment zd for dimension d from a Chinese restaurant process with hyperparameter α (with α from a vague hyperprior).
2. For each group g in the dimension partition {zd}: - (a) For each sampled datapoint (or row) rεR, generate a cluster assignment zr9 from a Chinese restaurant process with hyperparameter αg (with αg from a vague hyperprior).
- (b) For each cluster c in the row partition for this group of dimensions {zdg}:
- i. For each dimension d, generate component model parameters {right arrow over (θ)}cd from an appropriate prior and {right arrow over (λ)}d.
- ii. For each data cell x(r,d) in this component zrx
d =c for dεD), generate its value from an appropriate likelihood and {right arrow over (θ)}cd.
The model encodes a very different inductive bias than the IBP, discovering independent systems of categories over heterogeneous data vectors, as opposed to features that are typically additively combined. It is also instructive to contrast the asymptotic capacity of our model with that of a Dirichlet process mixture. The DP mixture has arbitrarily large asymptotic capacity as the number of samples goes to infinity. Put differently, it can model any distribution over finite dimensional vectors given enough data. However, if the number of dimensions (or features) is taken to infinity, it is no longer asymptotically consistent: if we generate a sequence of datasets by sampling the first K1 dimensions from a mixture and then append K2>>K1 dimensions that are constant valued (e.g. the price of tea in China), it will eventually be forced to model only those dimensions, ignoring the statistical structure in the first K1. In contrast, the model has asymptotic capacity both in terms of the number of samples and the number of dimensions, and is infinitely exchangeable with respect to both quantities.
As a consequence, it is self-consistent over the subset of variables measured, and can thus enjoy considerable robustness in the face of noisy, missing, and irrelevant measurements or confounding statistical signals. This should be especially helpful in demographic settings and in high-throughput biology, where noisy, or coherently co-varying but orthogonal, measurements are the norm, and each data vector arises from multiple, independent generative processes in the world.
The algorithm builds upon a general-purpose MCMC algorithm for probabilistic programs and specializes three of the kernels. It scales linearly per iteration in the number of rows and columns and includes inference over all hyperparameters.
A preliminary comparison of the learning curves for cross-categorization and one-versus-all SVMs on synthetic 5-class classification, averaged over datasets generated from 10 dimensional Bernoulli mixtures.
The detailed mechanisms by which the latent variables introduced by this method above those in a regular DP mixture improve mixing performance. Massively parallel implementations exploit the conditional independencies in the described model. Because the described method is essentially parameter free (e.g. with improper uniform hyperpriors), robust to noisy and/or irrelevant measurements generated by multiple interacting causes, and supports arbitrarily sparsely observed, heterogeneous data, it may be broadly applicable in exploratory data analysis. Additionally, the performance of our MCMC algorithm suggests that the described approach to nesting latent variable models in a Dirichlet process over dimensions may be applied to generate robust, rapidly converging, cross-cutting variants of a wide variety of nonparametric Bayesian techniques.
A co-assignment matrix for dimensions, where:
Cij=Pr[zi=zj]
That is, the probability that dimensions i and j share a common cause and therefore are modeled by the same Dirichlet process mixture. Labels show the consensus dimension groups (probability >0:75). These reflect attributes that share a common cause and thus co-vary, while the remainder of the matrix captures correlations between these discovered causes, for instance, mammals rarely have feathers or fly, ungulates are not predators, and so forth. Each dimension group picks out a different cross-cutting categorization of the rows (e.g. vertebrates, birds, canines, . . . ; not shown).
An example may include data for 4273 hospitals by 74 variables, including quality scores and various spending measurements. The data is analyzed (˜1 hour for convergence) with no preprocessing or missing data imputation.
Each box contains one consensus dimension group and the number of categories according to that group. In accordance with custom statistical analyses, no causal dependence between quality of care, hospital capacity, and spending is found, though each kind of measurement results in a different clustering of the hospitals. Also recovered is the cost structure of modern hospitals (e.g. increased long term care causes increased ambulance costs, likely due to an increase in at-home mishaps). Standard clustering methods miss most of this type cross-cutting structure.
Veritable and associated Veritable APIs make use of a predictive database that finds the causes behind data and uses these causes to predict and explain the future in a highly automated fashion heretofore unavailable, thus allowing any developer to carry out scientific inquires against a dataset without requiring custom programming and consultation with mathematicians and other such experts.
Veritable works by searching through the massive hypothesis space of all possible relationships present in a dataset, using an advanced Bayesian machine learning algorithm. The described Veritable technologies offer developers: state of the art inference performance and predictive accuracy on a very wide range of real-world datasets, with no manual parameter tuning whatsoever; scalability to very large datasets, including very high-dimensional data with hundreds of thousands or millions of columns; completely flexible predictions (e.g., predict the value of any subset of columns, given values for any other subset) without any retraining or adjustment; and quantification of the uncertainty associated with its predictions, since the system is built around a fully Bayesian probability model.
Described applications built on top of Veritable range from predicting heart disease, to understanding health care expenditures, to assessing business opportunities and scoring a likelihood to successfully “close” such business opportunities (e.g., to successfully commensurate a sale, contract, etc.).
Consider the problems with real-world data There's different kinds of data mixed together. Data is entered and maintained by people with other things on their minds and real work to get done. Real-world data contains errors and blanks, such as null values in places where populated values are appropriate. Users of the data may be measuring the wrong thing, or may be measuring the same thing in ten different ways. And in many organizations, no one knows precisely what data is contained in the database. Perhaps there was never a DBA (Data Base Administrator) for the organization, or the DBA left, and ten years of sedimentary layers of data has since built up. All of these are very realistic and common problems with “real-world” data found in production databases for various organizations, in contrast to pristine and small datasets that may be found in a laboratory or test setting.
A system is needed that can make sense of data as it exists in real businesses and does not require an pristine dataset or conform to typical platonic ideals of what data should look like. A system is needed that can be queried for many different questions about many different variables, in real time. A system is needed which is capable of getting at the hidden structures in such data, that is, which variables matter and what are the segments or groups within the data. At the same time, the system must be trustworthy, that is, it can't lie to the users by providing erroneous relationships and predictions. Such a system shouldn't reveal things that aren't true and shouldn't report ghost patterns may exist in a first dataset, but won't hold up overall. Such desirable characteristics are exceedingly difficult to attain with customized statistical analysis and customized predictive modeling, and wholly unheard of in automated systems.
According to the described embodiments, the resulting database appears to its users much like a traditional database. But instead of selecting columns from existing rows, users may issue predictive query requests via a structured query language. Such a structured language, rather than SQL may be referred to as Predictive Query Language (“PreQL”). PreQL is not to be confused with PQL which is short for the “Program Query Language.”
PreQL is thus used to issue queries against the database to predict values. Such a PreQL query offers the same flexibility as SQL-style queries. When exploring structure, users may issue PreQL queries seeking notions of similarity that are hidden or latent in the overall data without advanced knowledge of what those similarities may be. When used in a multi-tenant database system against a massive cloud based database and its dataset, such features are potentially transformative in the computing arts.
According to certain embodiments, Veritable utilizes a specially customized probabilistic model based upon foundational CrossCat modeling. CrossCat is a good start but could nevertheless be improved. For instance, it was not possible to run equations with CrossCat, which is solved via the core of Veritable which uses a particular engine implementation enabling such equation execution. Additionally, prior models matched data with the model to understand hidden structure, like building a probabilistic index, but was so complex that its users literally required advanced mathematics and probability theory understanding simply to implement the models for any given dataset, rendering mere mortals incapable of realistically using such models. Veritable implementations described herein provide a service which includes distributed processing, job scheduling, persistence, check-pointing, and a user-friendly API or front-end interface which accepts users' questions and queries via the PreQL query structure. Other specialized front end GUIs and interfaces are additionally described to solve for particular use cases on behalf of users and provide other simple interfaces to complex problems of probability.
What is probability? There are many perspectives, but probability may be described as a statement, by an observer, about a degree of belief in an event, past, present, or future. Timing doesn't matter
What is uncertainty? An observer, as noted above, doesn't know for sure whether an event will occur, notwithstanding the degree of belief in such an event having occurred, occurring, or to occur in the future.
Probabilities are assigned relative to knowledge or information context. Different observers can have different knowledge, and assign different probabilities to same event, or assign different probabilities even when both observers have the same knowledge. Probability, as used herein, is a number between “0” (zero) and “1” (one), in which 0 means the event is sure to not occur on one extreme of a continuum and where 1 means the event is sure to occur on the other extreme of the same continuum. Both extremes are uninteresting because they represent a complete absence of uncertainty.
A probability ties belief to one event. A probability distribution ties beliefs to every possible event, or at least, every event we want to consider. Choosing the outcome space is an important modeling decision. Summed over all outcomes in space is total probability which must be a total of “1,” that is to say, one of the outcomes must occur. Probability distributions are convenient mathematical forms that help summarize the system's beliefs in the various probabilities, but choosing a standard distribution is a modeling choice in which all models are wrong, but some are useful.
Consider for example, a Poisson distribution which is a good model when some event can occur 0 or more times in a span of time. The outcome space is the number of times the event occurs. The Poisson distribution has a single parameter, which is the rate−the average number of times. Its mathematical form has some nice properties: Defined for all the non-negative integers Sums to 1.
Many examples exist besides the Poisson distribution. Each such standard distribution encompasses a certain set of assumptions, such as a particular outcome space, a particular way of assigning probabilities to outcomes, etc. If you work with them, you'll start to understand why some are nice and some are frustrating if not outright evil.
Veritable utilizes distributions which move beyond the standard distributions with specially customized modeling thus allowing for a more complex outcome space and further allowing for more complex ways of assigning probabilities to outcomes. Depicted at slide 8B above is what's called a mixture distribution combining a bunch of simpler distributions to form a more complex one. A mixture of Gaussians to model any distribution may be employed, while still assigning probabilities to outcomes, yielding a more involved mathematical relationship.
With more complex outcome spaces, a Mondrian process defines a distribution on k-dimensional trees, providing means for dividing up a square or a cube. The outcome space is all possible trees and resulting divisions look like the famous painting. The outcome space is more structured than what is offered by the standard distributions. CrossCat does not use the Mondrian process, but it does use a structured outcome space. Veritable utilizes the Mondrian process in select embodiments.
At a high level, probability theory is a generalization of logic. Just like computers can use logic to reason deductively, probability lets computer reason inductively, generalize, categorize, etc. Probability gives us a way to combine different sources of information in a systematic manner, that is, utilizing automated computer implemented functionality, even when that information is vague, or uncertain, or ambiguous.
Such assumptions don't just give us a hypothesis space, they also give us a way of assigning probabilities to them, yielding a probability distribution on hypotheses, given the actual data observed.
There can be a great many hypotheses and finding the best ones to explain the data is not a straight forward or obvious proposition.
Modeling makes assumptions and using the assumptions defines a hypothesis space. Probabilities are assigned to the hypotheses given data observed and then inference is used to figure out which of those explanatory hypotheses is the best or are plausible.
Many approaches exist and experts in the field do not agree on how to select the best hypothesis. In simple cases, we can use math to solve the equations directly. Optimization methods are popular such as hill climbing and its relatives. Veritable may use any such approaches, but according to certain described embodiments, Monte Carlo methods are specifically utilized in which a random walk is taken through the space of hypotheses. Random doesn't mean stupid, of course. In fact, efficiently navigating these huge spaces is a one of the innovations utilized to improve the path taken by the random walk.
According to exemplary models, tabular data is provided such that rows are individual entities and columns contain a certain piece of information about the respective entity (e.g., field). Different kinds of columns in one table are acceptable.
Utilizing cross-categorization, columns/features are grouped into views and rows/entities group into categories. View 3//Category 3 from slide 19 above is expanded to the right having 8 features and 4 entities. The highlighted rows are in different categories in View 3 but are in the same category in another view. Zoom in again, and it is seen that this category contains the actual data points corresponding to the cell values in the table for just the columns in this view, and just the rows in this category.
A single cross-categorization is a particular way of slicing and dicing the table. First by column and then by row. It's a particular kind of process to yield a desired structured space. Utilizing concepts discussed above with respect to probability distributions, a probability is then assigned to each cross-categorization. More complex cross-categorizations yielding more views and more categories are less probable in and of themselves and typically are warranted only when the data really supports them.
These numbers are so massive that it may be helpful to consider the following context. The number in red, that is, the horizontal line near the very bottom of the plot on the right, is the total number of web pages. Thus, Google only needs to search through the 17th bell number or so. The space is so unimaginably massive that it simply is not possible to explore it exhaustively. Moreover, because it's not smooth or concave, you can't just climb the hill either.
So where does the data come into cross-categorizations? Views pick out some of the columns and categories pick out some of the rows. Each column contains a single kind of data so each vertical strip within a category contains typed data such as numerical, categorical, etc. Now, the basic standardized distributions may be utilized more effectively. In certain embodiments, each collection of points is modeled with a single simple distribution. Basic distributions are pre-selected which work well for each data type and each selected basic distribution is only responsible for explaining a small subset of the actual data, for which it is particularly useful. Then using the mixture distribution discussed above, the basic distributions are combined such that a bunch of simple distributions are used to make a more complex one. The structure of the cross-categorization is used to chop up the data table into a bunch of pieces and each piece is modeled using the simple distribution selected of the data type, yielding a big mixture distribution of the data.
So what does this look like? Go back to our mammals example providing a sample cross-categorization having two views. The view on the right has the habitat and feeding style columns and the rows are divided into four categories land mammals (Persian cat through Zebra), sea predators (dolphin through walrus), baleen whales (blue whale and humpback whale only), and the outlier amphibious beaver (e.g., both land and water living; we do not suggest that mammal beavers have gills). The view on the left has another division in which the primates are grouped together, large mammals are grouped, grazers are grouped, and then a couple of data oddities at the bottom (bat and seal). Even with a small dataset it is easy to imagine different ways of dividing the data up. But data is ambiguous. There is no perfect or obviously right division. For all the groupings that seemingly fit correctly, certain groupings may seem awkward or poor fitting. The systematic process of applying various models and assumptions makes tradeoffs and compromises, which is why even experts cannot agree on a single approach. Nevertheless, the means described herein permits use of a variety of available models such that these tradeoffs and compromises may be exploited to further benefit the system.
Results are thus not limited to a single cross-categorization. Instead, a collection of them are utilized and such a collection when used together tells us about the hidden structure of the data. For instance, if they're all the same, then there was no ambiguity in the data, but such a result doesn't occur with real-world data, despite being a theoretical possibility. Conversely, if they're all completely different, that means we couldn't find any structure in the data, which sometimes happens, and requires some additional post-processing to get at the uncertainty, such as feeding in additional noise. Typically, however, something in between occurs, and some interesting hidden structure is revealed from the data.
The specially customized cross-categorization implementation represents the core of Veritable. This core is not directly exposed to the users who interface via APIs, PreQL, and specialized utility GUIs and interfaces, but such users nevertheless benefit from the functionality which drives these other capabilities.
The Veritable core utilizes Monte Carlo methods for certain embodiments.
First, you need to get data into the system so API calls are provided to upload data into the system. A row is the basic unit. An API call for “Analyze” kicks off a learning pass applying the specially customized CrossCat model for the uploaded data. It's also possible to specify an existing dataset, or to define a sub-set of data from a larger dataset.
Cross-categorizations are found in the data that are most plausible explanations of the data at hand. Though such functions happy in the background out of the eyes of the user, such functionality is computationally intensive and is thus, well suited for a distributed computing structure provided by a cloud based multi-tenant database system architecture.
Using PreQL, specialized queries are thus made feasible. For instance, we can ask: for a given column, what are the other columns that are predictively related to it? In terms of the cross-categorizations, we tabulate how often each of the other columns appears in the same view as the input column, thus revealing what matters and what doesn't matter.
Using PreQL, we can ask what rows “go together.” Such a feature can be conceptualized as clustering, except that there's more than one way to cluster. Consider the mammals example in which we additionally input a column via the PreQL query. The groups that are returned are in the context of that column.
Using PreQL, we can ask which rows are most similar to a given row. Rows can be similar in one context but dissimilar in another. For instance, killer whales and blue whales are a lot alike in some respects, but very different in others. The input column disambiguates.
Using PreQL, we can predict, that is, ask the system to render a prediction. With the cross-categorizations a prediction request is treated as a new row and we assign that row to categories in each cross-categorization. Then using the basic standardized distributions for each category the values we want to predict are predicted. Unlike conventional predictive analytics, the system provides for flexible predictive queries thus allowing the user of the PreQL query to specify as many or as few columns as they desire and thus allowing the system to predict as many or as few as the user wants.
For instance, consider classification or regression in which all but one of the columns are used to predict a single target column. Veritable's core, via the APIs, can predict using a single target column or can predict using a few target columns at the user's discretion. For instance, a user can query the system asking: will an opportunity close AND at what amount? Such capabilities do not exist using conventional means.
Alternatively, the user can take an incomplete row and predict all of the missing values to fill in the blanks. At the extreme, the user can begin a table with many missing values and render a table where all of the blanks were filled in. Specialized tools for this particular use case are discussed below in which functionality allows the user to trade off confidence for more or less data, such that more data (or all the data) can be populated with degrading confidence or only some data is populated, above a given confidence, and so forth. A specialized GUI is additionally provided and described for this particular use case. Such a GUI calls the predict query via PreQL via an API on behalf of the user, but fundamentally exercises Veritable's core.
Here the table is provided as being 61% filled. No values are predicted, but the user may simply move the slider to increase the data fill for the missing values, causing the GUI's functionality to utilize the predict function on behalf of the user.
Here the table is provided as being 73% filled. Some but not all values are predicted. Not depicted here is the confidence threshold which is hidden from the user. Alternative interfaces allow the user to specify such a threshold.
Here the table is provided as being 100% filled. All values are predicted, but it may be necessary to degrade the confidence somewhat to attain 100% fill. Though such a fill may nevertheless be feasible at acceptable levels of confidence. In each of the instances, the grey scale values show the original data and the blue values depict the predicted values which do not actually exist in the underlying table. In certain embodiments, the chosen fill level, selected by the user via the slider bar, can be “saved” to the original or a copy of the table, thus resulting in the predictive values provided being saved or input to the cell locations. Meta data can be used to recognize later that such values were predicted and not sourced.
Other specialized GUIs and API tools include business opportunity scoring, next best offer, etc. These and other are described in additional detail below.
When making predictions, it is helpful to additionally let the users know whether they can trust the result. That is, how confident is the result and is the system literally capable of saying: “I do not know.” With such a system, the result may come back and tell the user if the answer is 1 or between 1 and 10 or between infinity and infinity.
With probabilities, the system can advise the user that it is 90% confident that the answer given is real, accurate, and correct, or the system may alternatively return a result indicating that it simply lacks sufficient data, and thus, there is not enough known to render a prediction. For example, the Veritable core may be used to a complete data set as if it was real by filling in the missing but predicted information into a spreadsheet, from which the completed data may be used to draw conclusions, as is depicted in the above slides 33, 34, and 35. The control slider is feasible because when you complete income, for example, what is actually returned to the GUI functionality making the call is the respective persons' income distribution.
By using such a GUI interface or such a concept in general, the user is given control over accuracy and confidence. In such a way, the user can manipulate how much data is to be filled in and to what extent the confidence level applies. What Veritable does behind the scenes is to take a table of data with a bunch of typed columns, and then the depicted Perceptible GUI interface at slides 33, 34, and 35 asks for a prediction for every single cell. Then for each cell that is missing, the Perceptible GUI gets a distribution in return from the API call to the Veritable core for the individual cell. Then when the slider is manipulated by a user, functionality for the slider looks at the distributions and looks at their variances and then gives the estimates. Thus, for any given cell having a predicted result in place of the missing null value, having seen “a” and “b” and “c” it then returns a value for that column.
Starting with nothing more than raw data in a tabular form, such as data on paper, in a spreadsheet, or in tables of a relational database, an API call first made to upload or insert the data into the predictive database upon which Veritable operates and then an API call is made to the Veritable core to analyze the data. Upon insert, the data it looks just like all other data. But once uploaded and the analyze operation is initiated, a probabilistic model is executed against the data. So the Veritable core starts to look at the ways that the rows and the columns can interact with each other and start to build the various relationships and causations. A generated statistical index figures out how and which columns are related to another. Veritable goes through and says, for instance, these particular columns are likely to share a causal origin. The difficult problem is that Veritable must perform this analysis using real world realities rather than pristine and perfect datasets. With data that exists in the real world, some columns are junk, some columns are duplicates, some columns are heterogeneous, some columns are noisy with only sparse data, and Veritable's core functionality implementing the statistical index must be able to pull the appropriate relationships and causations out despite having to perform its analysis operations against real-world data.
There will also be no one right answer, as there are uncertainties. So Veritable does not just build up one statistical model but rather, Veritable builds multiple statistical indices as a distribution or an ensemble of statistical indices. Veritable performs its analysis by searching through a large space for all of the ways that the data provided can possibly interact.
The distribution of indices results in a model that is stored and is queryable by PreQL structured queries via Veritable's APIs. What Veritable figures out via the analysis operations is first how the columns group together and then how the various rows group together. The analysis thus discovers the hidden structure in the data to provide a reduced representation of a table that explains how rows and columns may be related such that they can be queried via PreQL.
PreQL structured queries allow access to the queryable model and its ensemble of indices through specialized calls, including: “RELATED,” “SIMILAR,” “GROUP,” AND “PREDICT,” each of which are introduced above at slides 28 through 32.
Beginning with PREDICT, calling an appropriate API for the PREDICT functionality enables users to predict any chosen sub-set of data predict any column/value. It is not required that an entire dataset be utilized to predict only a single value, as is typical with custom implemented models.
Using PREDICT, the user provides or fixes the value of any column and then the PREDICT API accepts the fixed values and those you want to predict. The functionality then queries the Veritable core asking: “Given a row that has these values fixed, as provided by the user, then what would the distribution be?” For instance, the functionality could fix all but one and column in the dataset and then predict the last one, as is done with customized models. But the PREDICT functionality is far more flexible, and thus, the user can change the column to be predicted at a whim and custom implemented models simply lack this functionality as they lack the customized mathematical constructs to predict for such unforeseen columns or inquiries. That is to say, absent a particular function be pre-programmed, the models simply cannot perform this kind of varying query, for instance, for a user exploring data making multiple distinct queries or simply changing the column or columns to be predicted as business needs and the underlying data and data structures of the client organization change over time.
Perhaps also the user does not know all the columns to fix. For instance, perhaps the dataset knows a few things about one user but lots about another user. For instance, an ecommerce site may know little about a non-registered passerby user but knows lots of information about a registered user with a rich purchase history. In such an example, the PREDICT functionality permits fixing or filling in only the stuff that is known without having to require all the data for all users, as some of the data is known to be missing. In such a way, the PREDICT functionality can still predict missing data elements with what is actually known.
Another capability using the PREDICT functionality is to specify or fix all the data in a dataset that is known, that is, non-null, and then fill in everything else. In such a way, a user can say that what is known in the dataset is known, but much data is understood to be missing, but render predictions for the data nevertheless. The PREDICT operation would thus increase the population of predicted data for missing or null-values by accepting decreasing confidence, until the all or a specified population percentage of data is reached, much like the Perceptible GUI and slider examples described above.
Another functionality using PREDICT is to fill in an empty set. So maybe data is wholly missing, and then you start generating data that represents new rows and the new data in those rows represents plausible data, albeit synthetic data.
In other embodiments, PREDICT can be used to populate data elements that are not known but should be present or may be present, yet are not filled in within the data set, thus allowing the PREDICT functionality to populate such data elements.
Another example is to use PREDICT to attain a certainty or uncertainty for any element and to display or return the range of plausible values or the element.
Next is the RELATED functionality. Given a table with columns or variables in it, Veritable's analysis behind the scenes divides the columns or variables into groups and because of the distributions there is more than one way to divide these columns or variables up. Take height for example. Giving the height column to an API call for RELATED, a user can query: “How confident can I be about the probability of the relationship existing in all the other columns with the height column so specified.” Then what is returned from the Veritable core for height is a confidence for every other column in the dataset which was not specified. So for example, the RELATED functionality may return for confidence to the height column, “Weight=1.0,” meaning that Veritable, according to the dataset, is extremely confident that there is a relationship between weight and height. Such a result is somewhat intuitive and expected. But other results may be less intuitive and thus provide interesting results for exploration and additional investigation. Continuing with the “height” example for the specified column to a RELATED API call, Veritable may return “Age=0.8” meaning that Veritable is quite sure, but not perfectly certain, due to, for instance, noisy data which precludes an absolute positive result. Perhaps also returned for the specified “height” column is “hair color=0.1” meaning there is realistically no correlation whatsoever between a person's height and their hair color, according to the dataset utilized. Thus, the RELATED functionality permits a user to query for what matters for a given column, such as the height column, and the functionality returns all the columns with a scoring of how related the columns are to the specified column, based on their probability.
Next is the SIMILAR functionality. Like the RELATED functionality, an API call to Veritable for SIMILAR accepts a row and then returns what other rows are most similar to the row specified Like the RELATED examples, the SIMILAR functionality returns the probability that a row specified and any respective returned row actually exhibits similarity. For instance, rather than specifying column, you specify “Fred” as a row in the dataset. Then you ask using the SIMILAR functionality, for “Fred,” what rows are scored based on probability to be the most like “Fred.” The API call can return all rows scored from the dataset or return only rows above or below a specified threshold. For instance, perhaps rows above 0.8 are the most interesting or the rows below 0.2 are most interesting, or both, or a range. Regardless, SIMILAR scores every row for the specified row and returns the rows and the score based on probably according to the user's constraints or the constraints of an implementing GUI, if any such constraints are given. Because the Veritable system figures out these relationships using its own analysis, there is more than way to evaluate for this inquiry. Thus, user must provide to an API call for SIMILAR the specified row to find and additionally the COLUMN which provides how you the user constructing the PreQL query or API call actually cares about the data. Thus, the API call requires both row and column to be fixed. In such a way, providing, specifying, or fixing the column variable provides disambiguation information to Veritable and the column indication tells the Veritable core where to enter the index. Otherwise there would be too many possible ways to score the returned rows as the Veritable core could not disambiguate how the caller cares about the information for which a similarity is sought.
Next is the GROUP functionality. Sometimes rows tend to group up on noisy elements in the dataset when the Veritable core applies its analysis; yet these elements may result in groupings that are not actually important. We know that each column will appear in exactly one of the groups as a view and so Veritable permits using that column to identify the particular “view” that will be utilized. The GROUP functionality therefore implements a row centric operation like the SIMILAR functionality, but in contrast to an API call for SIMILAR where you must give a row and the SIMILAR call returns back a list of other rows and a score based on their probabilities of being related, with the GROUP functionality, the API call requires no row to be given or fixed whatsoever. Only a column is thus provided when making a call to the GROUP functionality.
Calling the GROUP functionality with a specified or fixed column causes the functionality to return the groupings of the ROWS that seem to be related or correlated in some way based on Veritable's analysis.
In such a way, use of the PreQL structure queries permits programmatic queries into the predictive database in a manner similar to a programmer making SQL queries into a relational database. Rather than a “select” statement in the query the term is replaced with the “predict” or “similar” or “related” or “group” statements. For instance, an exemplary PreQL statement may read as follows: “PREDICT IS_WON, DOLLAR_AMOUNT FROM OPPORTUNITY WHERE STAGE=‘QUOTE’.” So in this example, “QUOTE” is the fixed column, “FROM” is the dataset from which an opportunity is to be predicted, the “PREDICT” term is the call into the appropriate function, “IS_WON” is the value to be predicted, that is to say, the functionality is to predict whether a given opportunity is likely or unlikely to be won where the “IS_WON” may have completed data for some rows but be missing for other rows due to, for example, pending or speculative opportunities, etc. “DOLLAR_AMOUNT” is the fixed value.
In certain embodiments, the above query is implemented via a specialized GUI interface which accepts inputs from a user via the GUI interface and constructs, calls, and returns data via the PREDICT functionality on behalf of the user without requiring the user actually write or even be aware of the underlying PreQL structure query made to the Veritable core.
There are additionally provide Veritable use case implementation embodiments and customized GUIs. According to a first embodiment, a specialized GUI implementation enables users to filter on a historical value by comparing a historical value versus a current value in a multi-tenant system. Filtering for historical data using a GUI's field option wherein the GUI displays current fields related to historical fields as is depicted.
Embodiments provide for the ability to filter historical data by comparing historical value versus a constant in a multi-tenant system. The embodiments utilize the Veritable core by calling the appropriate APIs to make queries on behalf of the GUI users. The GUI performs the query and then consumes the data which is then presented back to the end users via the interface. Consider for example, a sales person looking at the sales information in a particular data set. The interface can take the distributions provided by Veritable's core and produce a visual indication for ranking the information according to a variety of customized solutions and use cases.
For instance, in a particular embodiment, systems and methods for determining the likelihood of an opportunity to close using only closed opportunities is provided.
SalesCloud is an industry leading CRM application currently used by 125,000 enterprise customers. Customers see the value of storing the data in the Cloud. These customers appreciate a web based interface to view and act on their data, and these customers like to use report and dashboard mechanisms provided by the cloud based service. Presenting these various GUIs as tabs enables salespeople and other end users to explore their underlying dataset in a variety of ways to learn how their business is performing in real-time. These users also rely upon partners to extend the provided cloud based service capabilities through APIs.
A cloud based service that offers customers the opportunity to learn from the past and draw data driven insights is highly desirable as such functionality should help these customers make intelligent decisions about the future for their business based on their existing dataset.
The customized GUIs utilize Veritable's core to implement predictive models which may vary per customer organization or be tailored to a particular organizations needs via programmatic parameters and settings exposed to the customer organization to alter the configuration and operation of Veritable's functionality.
For instance, a GUI may be provided to compute and assign an opportunity score based on probability for a given opportunity reflecting the likelihood of that opportunity to close as a win or loss. The data set to compute this score would consists of all the opportunities that have been closed (either won/loss) in a given period of time, such as 1, 2, or 3 years or a lifetime of an organization, etc. Additional data elements from the customer organization's dataset may also be utilized, such as the account object as an input. Machine learning techniques implemented via Veritable's core, such as SVN, Regression, Decision Trees, PGM, etc., are then used to build an appropriate model to render the opportunity score and then the GUI depicts the information to the end user via the interface.
Systems and methods for determining the likelihood of an opportunity to close using historical trending data is additionally disclosed. For instance, a historical selector for picking relative or absolute dates is described.
With this example we enable users to look at how an opportunity has changed over time, independent of stage, etc. The user can additionally look at how that opportunity has matured from when created until when it was closed.
Systems and methods for determining the likelihood of an opportunity to close at a given stage using historical trending data. Where the above example operates independent of stage of the sales opportunity this example further focuses on the probability of closing at a given stage as a further limiting condition for the closure. Thus, customers are enabled to use the historical trending data to know exactly when the stage has changed and then additionally predict what factors were involved to move from stage 1 to 2, from stage 2 to 3 and so forth.
Systems and methods for determining the likelihood for an opportunity to close given social and marketing data is additionally disclosed. In this example, the dataset of the customer organization or whomever is utilizing the system is expanded on behalf of the end user beyond that which is specified and then that additionally information is utilized to further influence and educate the predictive models. For instance, certain embodiments pull information from an exemplary website such as “data.com,” and then the data is associated with each opportunity in the original data set to discover further relationships, causations, and hidden structure which can then be presented to the end user. Other data sources are equally feasible, such as pulling data from social networking sites, search engines, data aggregation service providers, etc.
In one embodiment, social data is retrieved and a sentiment is provided to the end-user via the GUI to depict how the given product is viewed by others in a social context. Thus, a salesperson can look at the persons linked in profile and with information from data.com or other sources the salesperson can additionally be given sentiment analysis in terms of social context for the person that the salesperson is actually trying to sell to. For instance, has the target purchaser commented about other products or have they complained about any other products, etc. Each of these data points and others based may help influence the model employed by Veritable's core to render a prediction.
Systems and methods for determining the likelihood for an opportunity to close given industry specific data is additionally disclosed. For instance, rather than using socially relevant data for social context of sentiment analysis, industry specific data can be retrieved and input to the predictive database upon which Veritable performs its analysis as described above, and from which further exploration can then be conducted by users of the dataset now having the industry specific data integrated therein.
According to other embodiments, datasets are explored beyond the boundaries of any particular customer organization having data within the multi-tenant database system. For instance, in certain embodiments, benchmark predictive scores are generated based on industry specific learning using cross-organizational data stored within the multi-tenant database system. For example, data mining may be performed against telecom specific customer datasets, given their authorization or license to do so. Such cross-organization data to render a much larger multi-tenant dataset can then be analyzed via Veritable and provide insights, relationships, causations, and additional hidden structure that may not be present within a single customer organizations' dataset. For instance, if as a customer you are trying to close a $100 k deal in the NY-NJ-Virginia tri-city area, the probability for that deal to close in 3 months may be, according to such analysis, 50% because past transactions have shown that it could take up to six months to close a $100 k telecom deal in NY-NJ-Virginia tri-city area when viewed in the context of multiple customer organizations' datasets. Many of the insights realized through such a process may be non-intuitive, yet capable of realization through application of the techniques described herein.
With industry specific data present within a given dataset it is possible to delve even deeper into the data and identify benchmark using such data for a variety of varying domains across multiple different industries. For instance, based on such data predictive analysis may review that, in a given region it takes six months to sell sugar in the mid west and it takes three months to sell laptops in the east coast, and so forth.
Then if a new opportunity arises and a vendor is trying to, for example, sell watches in California, the vendor can utilize such information to gain a better understanding of the particular regional market based on the predictions and confidence levels given.
Provided functionality can additionally predict information for a vertical sector as well as for the region. When mining a customer organization's dataset a relationship may be discovered that where customers bought a, those customers also bought b. These kinds of matching relationships are useful, but can be further enhanced. For instance, using the predictive analysis of Veritable it is additionally possible to identify the set of factors that led to a particular opportunity score (e.g., a visualized presentation of such analysis).
Thus, the GUI presents a 42% opportunity at the user interface but when the user mousse over the opportunity score, the GUI then displays sub-detailed elements that make up that opportunity score. The GUI makes the necessary Veritable based API calls on behalf of the user such that an appropriate call is made to the predictive platform to pull the opportunity score and display that information to the user as well as the sub-detail relationships and causations considered relevant.
The GUI can additionally leverage the predict and analyze capabilities of Veritable which upon calling a predict function for a given opportunity will return data necessary to create a histogram for an opportunity. So not only can the user be given a score, but the user can additionally be given the factors and guidance on how to interpret the information provided and what to do with such information.
Moreover, as the end-users, such as salespersons, see the data and act upon it, a feedback loop is created through which further data is input into the predictive database upon which additional predictions and analysis are carried out in an adaptive manner. For example, as the Veritable core learns more about the data the underlying models may be refreshed on a monthly basis by re-performing the analysis of the dataset so as to re-calibrate the data using the new data obtained via the feedback loop.
Additionally disclosed are systems and methods to deliver a matrix report for historical data in a multi-tenant system. For instance, consider a summary view of a matrix report according to provided embodiments.
Systems and methods to deliver a matrix report for historical data in a multi-tenant system follow the established matrix report format familiar to salesforce.com customers but which is limited only to current data and cannot display historical data. With Veritable's capabilities the historical data can additionally be provided via the matrix reports.
In a summary view, users can see data in an aggregate fashion. Each stage may consist of multiple opportunities and each might be able to be duplicated because each might change according to the amount or according to the stage, etc. Thus, if a user is looking at the last four weeks, then one opportunity may change from $500 to $1500 and thus be duplicated.
The cloud computing architecture executes functionality which runs across all the data for all tenants. Thus, for any cases, leads, and opportunities, the database maintains a history object into which all of audit data is retained such that a full and rich history can later be provided to the user at their request to show the state of any event in the past, without corrupting the current state of the data. Thus, while the underlying data must be maintained in its correct state for the present moment, a user may nevertheless utilize the system to display the state of a particular opportunity as it stood last week, or as it transitioned through past quarter, and so forth.
All of the audit data from history objects for various categories of data is then aggregated into a historical trending entity which is a custom object that stores any kind of data. This object is then queried by the different historical report types across multiple tenants to retrieve the necessary audit trail data such that any event at any time in the past can be re-created for the sake of reporting, predictive analysis, and exploration. The historical audit data may additionally be subjected to the analysis capabilities of Veritable by including it within a historical dataset for the sake of providing further predictive capabilities.
The algorithms to provide historical reporting capabilities are applied across all the tenant data which is common within the historical trending data object and the interim opportunity history and lead history, etc.
Within the matrix report, the data can also be visualized using salesforce.com's charting engine as depicted by the waterfall diagram.
Systems and methods to deliver waterfall charts for historical data in a multi-tenant system are thus provided. For instance, on the x axis is the weekly snapshot for all the opportunities being worked. The amounts are changing up and down and then are also grouped by stages. The waterfall enables a user to look at two points in time and by defining opportunities between day one and day two. Alternatively, waterfall diagrams can be used to group all opportunities into different stages as in the example above which every opportunity is mapped according to its stage allowing a user to look into the past and understand what the timing is for these opportunities to actually come through to closure.
Historical data and the audit history saved to the historical trending data object are enabled through snapshots and field history. Using the historical trending data object the desired data can then be queried. The historical trending data object may be implemented as one table with indexes on the table from which any of the desired data can then be retrieved. The software modules implementing the various GUIs and use cases populate the depicted table using the opportunity data retrieved from the historical trending data object's table.
Additionally disclosed are systems and methods to deliver waterfall chart for historical data in a multi-tenant system and historical trending in a multi-tenant environment. Systems and methods for using a historical selector for picking relative or absolute dates are additionally described.
These specialized implementations enable users to identify how the data has changed on a day to day basis or week to week basis or over a month to month basis, etc. The users can therefore can see the data that is related to the user's opportunities not just for the present time, but with this feature, the users can identify opportunities based on a specified time such as absolute time or relative time, so that they can see how the opportunity has changed over time. In this embodiment, time as a dimension is used to then provide a decision tree for the customers to pick either absolute date or a range of dates. Within the date customers can pick an absolute date, such as Jan. 1, 2013 or a relative date such as the first day of the current month or first day of the last month, etc.
This solves the problem of a sales manager or sales person needing to see how the opportunity has changed today versus the first day of this month or last month. With this capability, the user can take a step back in time, thinking back where they were a week ago, or a month ago and identify the opportunity by creating a range of dates and displaying what opportunities were created during those dates.
Thus, a salesperson wanting such information may have had ten opportunities and on Feb. 1, 2013, the salesperson's target buyer expresses interest in a quote, so the stage changes from prospecting to quotation. Another target buyer, however, says they want to buy immediately, so the state changes from quotation to sale/charge/close. The functionality therefore provides a back end which implements a decision tree with the various dates that are created. The result is that the functionality can give the salesperson a view of all the opportunities that are closing in the month of January, or February, or within a given range, etc.
The query for dates is unique because it is necessary to traverse the decision tree to get to the date the user picks and then enabling the user to additionally pick the number of snap shots, from which the finalized result set is determined, for instance, from Feb. 1 to Feb. 6, 2013.
Additionally described is the ability to filter historical data by comparing historical values versus current values in a multi-tenant system as is shown.
Additionally enabled is the ability to filter historical data by comparing historical values versus a constant in a multi-tenant system, referred to as a historical selector. Based on the opportunity or report type, the customer has the ability to filter on historical data using a custom historical filter. The interface provides the ability for the customer to look at all of the filters on the left that they can use to restrict a value or a field, thus allowing customers to filter on historical column data for any given value. Thus, a customer may look at all of the open opportunities for a given month or filter the data set according to current column data rather than historical. Thus, for a given opportunity a user at the interface can fill out the amount, stage, close date, probability, forecast category, or other data elements and then as the salesperson speaks with the target buyer, the state is changed from prospecting to quoting, to negotiation based on the progress that is made with the target buyer, and eventually to a state of won/closed or lost, etc. So maybe the target buyer is trying to decrease the amount of the deal and the salesperson is trying to increase the amount. All of that data and state changes (e.g., a change in amount can be a state change within a given phase of the deal) and the information is stored in the historical opportunity object which provides the audit trail.
As the current data changes the data in the current tables change and thus, historical data is not accessible to the customer. But the audit trail is retained and so it can be retrieved. For instance, the GUI enables a user to go back 12 months according to one embodiment. Such historical data and audit data may be processed with granularity of one day, and thus, a salesperson can go back in time and view how the data has changed overtime with within the data set with the daily granular reporting.
Thus, for any given opportunity object with all the object history and the full audit trail with daily granular data the historical trending entity object is used to allow the tool to pull the information about how these opportunities changed over time for the salesperson. Such metrics would be useful to other disciplines also, such as a service manager running a call center who gets 100 cases from sales agents wants to know how to close those calls, etc. Likewise, when running a marketing campaign, is it being spent in California, or Tokyo, etc., the campaign managers will want to know how to close the various leads an opportunities as well as peer back into history to see how events influenced the results of past opportunities.
Additional detail with respect to applying customized filters to historical data is further depicted, as follows:
According to one embodiment, the historical trending entity object is implemented via a historical data schema in which history data is stored in a new table core.historical_entity_data as depicted at Table 1 below:
Indices utilized in the above Table 1 include: organization_id, key_prefix, historic_entity_data_id. PK includes: organization_id, key_prefix, system_modstamp. Unique, find, and snapshot for given date and parent record: organization_id, key_prefix, parent_id, valid_to, valid_from. Indices organization_id, key_prefix, valid_to facilitate data clean up. Such a table is additionally counted against storage requirements according to certain embodiments. Usage may be capped at 100. Alternatively, when available slots are running low, old slots may be cleaned. Historical data management, row limits, and statistics may be optionally utilized. For new history the system assumes an average 20 byte per column and 60 effective columns (50 effective data columns+PK+audit fields) for the new history table, and thus, row size is 1200 bytes. For row estimates the system assumes that historical trending will have usage patterns similar to entity history. Since historical trending storage is charged to the customer's applicable resource limits, it is expected that usage will not be heavier than usage of entity history.
Sampling of production data revealed recent grow in row count for entity history is ˜2.5 B (billion) rows/year. Since historical trending will store a single row for any number of changed fields, an additional factor of 0.78 can be applied. Since historical trending will only allow 4 standard and at most 5 custom objects, additional factor of 0.87 can be used to only include top standard and custom objects contributing to entity history. With additional factor of 0.7 to only include UE and EE organizations, the expected row count for historical trending is 1.2 B row/year in the worst case scenario.
Historical data may be stored by default for 2 years and the size of the table is expected to stay around 2.4 B rows. Custom value columns are to be handled by custom indexes similar to custom objects. To prevent unintentional abuse of the system, for example, by using automated scripts, each organization will have a history row limit for each object. Such a limit could be between approximately 1 and 5 million rows per object which is sufficient to cover storage of current data as well as history data based on analyzed usage patterns of production data with only very few organizations (3-5) occasionally having so many objects that they would hit the configurable limit. The customized table can be custom indexed to help query performance.
High level use cases for such historical based data in a dataset to be analyzed and subjected to Veritable's predictive analysis include: Propensity to Buy and Lead Scoring for sales representatives and marketing users. For instance, sales users often get leads from multiple sources (marketing, external, sales prospecting etc.) and often times, in any given quarter, they have more leads to follow up than they have time. Sales representatives often need guidance with key questions such as: which leads have the highest propensity to buy, what is the likelihood of a sale, what is the potential revenue impact if this lead is converted to an opportunity, what is the estimated sale cycle based on historical observations if this lead is converted to an opportunity, what is the lead score for each of their leads in their pipeline so that sales representatives can discover the high potential sales leads in their territories, and so forth. Sales representatives may seek to determine the top ten products each account will likely buy based on the predictive analysis and the deal sizes if they successfully close, the length of the deal cycle based on the historical trends of similar accounts, and so forth. When sales representatives act on these recommendations, they can broaden their pipeline and increase their chance to meet or exceed quota, thus improving sales productivity, business processes, prospecting, and lead qualification.
Additional use cases for such historical based data may further include: likelihood to close/win and opportunity scoring. For instance, sales representatives and sales managers may benefit from such data as they often have many few deals in their current pipeline and must juggle where to apply their time and attention in any month/quarter. As these sales professionals approach the end of the sales period, the pressure to meet their quota is of significant importance. Opportunity scoring can assist with ranking the opportunities in the pipeline based on the probability of such deals to close, thus improving the overall effectiveness of these sales professionals.
Data sources may include such data as: Comments, sales activities logged, standard field numbers for activities (e.g., events, log a call, tasks etc.), C-level customer contacts, decision maker contacts, close dates, standard field numbers for times the close date has pushed, opportunity competitors, standard field opportunities, competitive assessments, executive sponsorship, standard field sales team versus custom field sales team as well as the members of the respective teams, chatter feed and social network data for the individuals involved, executive sponsor involved in a deal, DSRs (Deal Support Requests), and other custom fields.
Historical based data can be useful to Veritable's predictive capabilities for generating metrics such as Next Likelihood Purchase (NLP) and opportunity whitespace for sales reps and sales managers. For instance, a sales rep or sales manager responsible for achieving quarterly sales targets will undoubtedly be interested in: which types of customers are buying which products; which prospects most resemble existing customers; are the right products being offered to the right customer at the right price; what more can we sell to my customer to increase the deal size, and so forth. Looking at historical data of things that similar customers have purchased to uncover selling trends, and using such metrics yields valuable insights to make predictions about what customers will buy next, thus improving sales productivity and business processes.
Another capability provided to end users is to provide customer references on behalf of sales professionals and other interested parties. When sales professions require customer references for potential new business leads they often spend significant time searching through and piecing together such information from CRM sources such as custom applications, intranet sites, or reference data captured in their databases. However, the Veritable core and associated use case GUIs can provide key information to these sales professions. For instance, the application can provide data including that is grouped according to industry, geography, size, similar product footprint, and so forth, as well as provide in one place what reference assets are available for those customer references, such as customer success stories, videos, best practices, which reference customers are available to chat with a potential buyer, customer reference information grouped according to the contact person's role, such as CIO, VP of sales, etc., which reference customers have been over utilized and thus may not be good candidate references at this time, who are the sales representatives or account representatives for those reference customers at the present time or at any time in the past, who is available internal to a organization to reach or make contact with the reference customer, and so forth. This type of information is normally present in database systems but is not organized in such a way and is extremely labor intensive to retrieve, however, Veritable's analysis can identify such relationships and hidden structure in the data which may then be retrieved and displayed by specialized GUI interfaces for end-users. Additionally, the functionality can identify the most ideal or the best possible reference customer among many based on predictive analysis which builds the details of a reference customer into a probability to win/close the opportunity, which is data wholly unavailable from conventional systems.
In other embodiments, filter elements are provided to the user to narrow or limit the search according to desire criteria, such as industry, geography, deal size, products in play etc. Such functionality thus aids sales professionals with improving sales productivity and business processes.
According to other embodiments, functionality is provided to predict forecast adjustments on behalf of sales professionals. For instance, businesses commonly have a system of sales forecasting as part of their critical management strategy. Yet, such forecasts are, by nature, inexact. The difficulty is knowing which direction such forecasts are wrong and then turning that understanding into an improved picture of how the business is doing. Veritable's functionality can improve such forecasting using existing data of a customer organization including existing forecasting data. For instance, applying Veritable's analysis to past forecast to the business can aid in trending and with improving existing forecasts into the future which have yet to be realized. Sales managers are often asked to provide their judgment or adjustment on forecasting data for their respective sales representatives which requires such sales managers to aggregate their respective sales representatives' individual forecasts. This is a labor intensive process which tends to induce error. Sales managers are intimately familiar with their representatives' deals and they spend time reviewing them on a periodic basis as part of a pipeline assessment. Improved forecasting results can aid such managers with rendering improved judgments and assessments as well as help with automating the aggregating function which is often carried out manually or using inefficient tools, such as spreadsheets, etc.
In such an embodiment, Veritable's analysis function mines past forecast trends by the sales representatives for relationships and causations such as forecast versus quota versus actual for the past eight quarters or other time period, and then provides a recommended judgment and/or adjustment that should be applied to the current forecast. By leveraging the analytical assessment at various levels of the forecast hierarchy, organizations can reduce the variance between individual sales representative's stipulated quotas, forecasts, and actuals, over a period of time, thereby narrowing deltas between forecast and realized sales via improved forecast accuracy.
Additional functionality enables use case GUI interfaces to render a likelihood to renew an opportunity or probability of retention for an opportunity by providing a retention score. Such functionality is helpful to sales professionals as such metrics can influence where a salesperson's time and resources are best spent so as to maximize revenue.
Unbounded Categorical Data types model categorical columns where new values that are not found in the dataset can show up. For example, most opportunities will be replacing one of a handful of common existing systems, such as an Oracle implementation, but a new opportunity might be replacing a new system which has not been seen in the data ever before.
The client devices 306A-C each individually transmit request packets 316 to the remote host organization 310 via the network 325. The host organization 310 may responsively send response packets 315 to the originating customer organization to be received via the respective client devices 306A-C. Such interactions thus establish the communications necessary to transmit and receive information in fulfillment of the described embodiments on behalf of each the customer organizations and the host organization 310 providing the cloud based computing services including access to the Veritable functionality described.
Within host organization 310 is a request interface 375 which receives the packet requests 315 and other requests from the client devices 306A-C and facilitates the return of response packets 316. Further depicted is a PreQL query interface 380 which operates to query the predictive database 350 in fulfillment of such request packets from the client devices 306A-C, for instance, issuing API calls for PreQL structure query terms such as “PREDICT,” “RELATED,” “SIMILAR,” and “GROUP.” Also available are the API calls for “UPLOAD” and “ANALYZE,” so as to upload new data sets or define datasets to the predictive database 350 and trigger the Veritable core 390 to instantiate analysis of such data. Server side application 385 may operate cooperatively with the various client devices 306A-C. Veritable core 390 includes the necessary functionality to implement the embodiments described herein.
Environment 210 is an environment in which an on-demand database service exists. User system 212 may be any machine or system that is used by a user to access a database user system. For example, any of user systems 212 can be a handheld computing device, a mobile phone, a laptop computer, a work station, and/or a network of computing devices. As illustrated in
An on-demand database service, such as system 216, is a database system that is made available to outside users that do not need to necessarily be concerned with building and/or maintaining the database system, but instead may be available for their use when the users need the database system (e.g., on the demand of the users). Some on-demand database services may store information from one or more tenants stored into tables of a common database image to form a multi-tenant database system (MTS). Accordingly, “on-demand database service 216” and “system 216” is used interchangeably herein. A database image may include one or more database objects. A relational database management system (RDMS) or the equivalent may execute storage and retrieval of information against the database object(s). Application platform 218 may be a framework that allows the applications of system 216 to run, such as the hardware and/or software, e.g., the operating system. In an embodiment, on-demand database service 216 may include an application platform 218 that enables creation, managing and executing one or more applications developed by the provider of the on-demand database service, users accessing the on-demand database service via user systems 212, or third party application developers accessing the on-demand database service via user systems 212.
The users of user systems 212 may differ in their respective capacities, and the capacity of a particular user system 212 might be entirely determined by permissions (permission levels) for the current user. For example, where a salesperson is using a particular user system 212 to interact with system 216, that user system has the capacities allotted to that salesperson. However, while an administrator is using that user system to interact with system 216, that user system has the capacities allotted to that administrator. In systems with a hierarchical role model, users at one permission level may have access to applications, data, and database information accessible by a lower permission level user, but may not have access to certain applications, database information, and data accessible by a user at a higher permission level. Thus, different users will have different capabilities with regard to accessing and modifying application and database information, depending on a user's security or permission level.
Network 214 is any network or combination of networks of devices that communicate with one another. For example, network 214 can be any one or any combination of a LAN (local area network), WAN (wide area network), telephone network, wireless network, point-to-point network, star network, token ring network, hub network, or other appropriate configuration. As the most common type of computer network in current use is a TCP/IP (Transfer Control Protocol and Internet Protocol) network, such as the global internetwork of networks often referred to as the “Internet” with a capital “I,” that network will be used in many of the examples herein. However, it is understood that the networks that the claimed embodiments may utilize are not so limited, although TCP/IP is a frequently implemented protocol.
User systems 212 might communicate with system 216 using TCP/IP and, at a higher network level, use other common Internet protocols to communicate, such as HTTP, FTP, AFS, WAP, etc. In an example where HTTP is used, user system 212 might include an HTTP client commonly referred to as a “browser” for sending and receiving HTTP messages to and from an HTTP server at system 216. Such an HTTP server might be implemented as the sole network interface between system 216 and network 214, but other techniques might be used as well or instead. In some implementations, the interface between system 216 and network 214 includes load sharing functionality, such as round-robin HTTP request distributors to balance loads and distribute incoming HTTP requests evenly over a plurality of servers. At least as for the users that are accessing that server, each of the plurality of servers has access to the MTS' data; however, other alternative configurations may be used instead.
In one embodiment, system 216, shown in
One arrangement for elements of system 216 is shown in
Several elements in the system shown in
According to one embodiment, each user system 212 and all of its components are operator configurable using applications, such as a browser, including computer code run using a central processing unit such as an Intel Pentium® processor or the like. Similarly, system 216 (and additional instances of an MTS, where more than one is present) and all of their components might be operator configurable using application(s) including computer code to run using a central processing unit such as processor system 217, which may include an Intel Pentium® processor or the like, and/or multiple processor units.
According to one embodiment, each system 216 is configured to provide webpages, forms, applications, data and media content to user (client) systems 212 to support the access by user systems 212 as tenants of system 216. As such, system 216 provides security mechanisms to keep each tenant's data separate unless the data is shared. If more than one MTS is used, they may be located in close proximity to one another (e.g., in a server farm located in a single building or campus), or they may be distributed at locations remote from one another (e.g., one or more servers located in city A and one or more servers located in city B). As used herein, each MTS may include one or more logically and/or physically connected servers distributed locally or across one or more geographic locations. Additionally, the term “server” is meant to include a computer system, including processing hardware and process space(s), and an associated storage system and database application (e.g., OODBMS or RDBMS) as is well known in the art. It is understood that “server system” and “server” are often used interchangeably herein. Similarly, the database object described herein can be implemented as single databases, a distributed database, a collection of distributed databases, a database with redundant online or offline backups or other redundancies, etc., and might include a distributed database or storage network and associated processing intelligence.
User system 212, network 214, system 216, tenant data storage 222, and system data storage 224 were discussed above in
Application platform 218 includes an application setup mechanism 338 that supports application developers' creation and management of applications, which may be saved as metadata into tenant data storage 222 by save routines 336 for execution by subscribers as one or more tenant process spaces 304 managed by tenant management process space 310 for example. Invocations to such applications may be coded using PL/SOQL 334 that provides a programming language style interface extension to API 332 (e.g., a PreQL or JSON API). Invocations to applications may be detected by one or more system processes, which manages retrieving application metadata 316 for the subscriber making the invocation and executing the metadata as an application in a virtual machine.
Each application server 300 may be communicably coupled to database systems, e.g., having access to system data 225 and tenant data 223, via a different network connection. For example, one application server 3001 might be coupled via the network 214 (e.g., the Internet), another application server 300N-1 might be coupled via a direct network link, and another application server 300N might be coupled by yet a different network connection. Transfer Control Protocol and Internet Protocol (TCP/IP) are typical protocols for communicating between application servers 300 and the database system. However, it will be apparent to one skilled in the art that other transport protocols may be used to optimize the system depending on the network interconnect used.
In certain embodiments, each application server 300 is configured to handle requests for any user associated with any organization that is a tenant. Because it is desirable to be able to add and remove application servers from the server pool at any time for any reason, there is preferably no server affinity for a user and/or organization to a specific application server 300. In one embodiment, therefore, an interface system implementing a load balancing function (e.g., an F5 Big-IP load balancer) is communicably coupled between the application servers 300 and the user systems 212 to distribute requests to the application servers 300. In one embodiment, the load balancer uses a least connections algorithm to route user requests to the application servers 300. Other examples of load balancing algorithms, such as round robin and observed response time, also can be used. For example, in certain embodiments, three consecutive requests from the same user may hit three different application servers 300, and three requests from different users may hit the same application server 300. In this manner, system 216 is multi-tenant, in which system 216 handles storage of, and access to, different objects, data and applications across disparate users and organizations.
As an example of storage, one tenant might be a company that employs a sales force where each salesperson uses system 216 to manage their sales process. Thus, a user might maintain contact data, leads data, customer follow-up data, performance data, goals and progress data, etc., all applicable to that user's personal sales process (e.g., in tenant data storage 222). In an example of a MTS arrangement, since all of the data and the applications to access, view, modify, report, transmit, calculate, etc., can be maintained and accessed by a user system having nothing more than network access, the user can manage his or her sales efforts and cycles from any of many different user systems. For example, if a salesperson is visiting a customer and the customer has Internet access in their lobby, the salesperson can obtain critical updates as to that customer while waiting for the customer to arrive in the lobby.
While each user's data might be separate from other users' data regardless of the employers of each user, some data might be organization-wide data shared or accessible by a plurality of users or all of the users for a given organization that is a tenant. Thus, there might be some data structures managed by system 216 that are allocated at the tenant level while other data structures might be managed at the user level. Because an MTS might support multiple tenants including possible competitors, the MTS may have security protocols that keep data, applications, and application use separate. Also, because many tenants may opt for access to an MTS rather than maintain their own system, redundancy, up-time, and backup are additional functions that may be implemented in the MTS. In addition to user-specific data and tenant specific data, system 216 might also maintain system level data usable by multiple tenants or other data. Such system level data might include industry reports, news, postings, and the like that are sharable among tenants.
In certain embodiments, user systems 212 (which may be client systems) communicate with application servers 300 to request and update system-level and tenant-level data from system 216 that may require sending one or more queries to tenant data storage 222 and/or system data storage 224. System 216 (e.g., an application server 300 in system 216) automatically generates one or more SQL statements or PreQL statements (e.g., one or more SQL or PreQL queries respectively) that are designed to access the desired information. System data storage 224 may generate query plans to access the requested data from the database.
Each database can generally be viewed as a collection of objects, such as a set of logical tables, containing data fitted into predefined categories. A “table” is one representation of a data object, and may be used herein to simplify the conceptual description of objects and custom objects as described herein. It is understood that “table” and “object” may be used interchangeably herein. Each table generally contains one or more data categories logically arranged as columns or fields in a viewable schema. Each row or record of a table contains an instance of data for each category defined by the fields. For example, a CRM database may include a table that describes a customer with fields for basic contact information such as name, address, phone number, fax number, etc. Another table might describe a purchase order, including fields for information such as customer, product, sale price, date, etc. In some multi-tenant database systems, standard entity tables might be provided for use by all tenants. For CRM database applications, such standard entities might include tables for Account, Contact, Lead, and Opportunity data, each containing pre-defined fields. It is understood that the word “entity” may also be used interchangeably herein with “object” and “table.”
In some multi-tenant database systems, tenants may be allowed to create and store custom objects, or they may be allowed to customize standard entities or objects, for example by creating custom fields for standard objects, including custom index fields. In certain embodiments, for example, all custom entity data rows are stored in a single multi-tenant physical table, which may contain multiple logical tables per organization. It is transparent to customers that their multiple “tables” are in fact stored in one large table or that their data may be stored in the same table as the data of other customers.
The exemplary computer system 400 includes a processor 402, a main memory 404 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc., static memory such as flash memory, static random access memory (SRAM), volatile but high-data rate RAM, etc.), and a secondary memory 418 (e.g., a persistent storage device including hard disk drives and a persistent database and/or a multi-tenant database implementation), which communicate with each other via a bus 430. Main memory 404 includes stored indices 424, an analysis engine 423, and a PreQL API 425. Main memory 404 and its sub-elements are operable in conjunction with processing logic 426 and processor 402 to perform the methodologies discussed herein. The computer system 400 may additionally or alternatively embody the server side elements as described above.
Processor 402 represents one or more general-purpose processing devices such as a microprocessor, central processing unit, or the like. More particularly, the processor 402 may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, processor implementing other instruction sets, or processors implementing a combination of instruction sets. Processor 402 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like. Processor 402 is configured to execute the processing logic 426 for performing the operations and functionality which is discussed herein.
The computer system 400 may further include a network interface card 408. The computer system 400 also may include a user interface 410 (such as a video display unit, a liquid crystal display (LCD), or a cathode ray tube (CRT)), an alphanumeric input device 412 (e.g., a keyboard), a cursor control device 414 (e.g., a mouse), and a signal generation device 416 (e.g., an integrated speaker). The computer system 400 may further include peripheral device 436 (e.g., wireless or wired communication devices, memory devices, storage devices, audio processing devices, video processing devices, etc.).
The secondary memory 418 may include a non-transitory machine-readable or computer readable storage medium 431 on which is stored one or more sets of instructions (e.g., software 422) embodying any one or more of the methodologies or functions described herein. The software 422 may also reside, completely or at least partially, within the main memory 404 and/or within the processor 402 during execution thereof by the computer system 400, the main memory 404 and the processor 402 also constituting machine-readable storage media. The software 422 may further be transmitted or received over a network 420 via the network interface card 408.
While the subject matter disclosed herein has been described by way of example and in terms of the specific embodiments, it is to be understood that the claimed embodiments are not limited to the explicitly enumerated embodiments disclosed. To the contrary, the disclosure is intended to cover various modifications and similar arrangements as are apparent to those skilled in the art. Therefore, the scope of the appended claims are to be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements. It is to be understood that the above description is intended to be illustrative, and not restrictive. Many other embodiments will be apparent to those of skill in the art upon reading and understanding the above description. The scope of the disclosed subject matter is therefore to be determined in reference to the appended claims, along with the full scope of equivalents to which such claims are entitled.
Claims
1. A method comprising:
- receiving a dataset in tabular form of columns and rows;
- triggering Veritable core analysis of the dataset received;
- identifying hidden structure of the dataset via the Veritable core analysis, the hidden structure including one or more of relationships and causations in the data for which such relationships and causations are not pre-defined by the dataset;
- storing the analyzed dataset including the hidden structure having the one or more of relationships and causations as a queryable model.
2. The method of claim 1, further comprising:
- querying the stored queryable model for predictive analysis.
3. The method of claim 1, further comprising one or more of:
- issuing a PreQL structure query against the queryable model, the PreQL structure comprising one of:
- a PREDICT term;
- a RELATED term;
- a SIMILAR term; and
- a GROUP term.
4. The method of claim 1, further comprising:
- providing a graphical user interface to the Veritable core as a cloud based computing service.
5. The method of claim 1, further comprising:
- providing a perceptible GUI as a cloud based service, the perceptible GUI accepting as input a data source within the predictive database;
- presenting at the perceptible GUI, a table representing the data source within the predictive database, wherein the table has a plurality of non-null values and a plurality of null values;
- providing a graphical slider mechanism at the perceptible GUI, wherein the graphical slider mechanism is manipulatable at a client device to increase and decrease a percentage of predictive fill for the null values of the table;
- populating null values of the table at the graphical perceptible GUI responsive to the graphical slider mechanism registering an increase in value to populate by the client device.
6. The method of claim 5, wherein populating the null values comprises:
- for every null value cell element in the data, retrieving a distribution via Veritable API calls from the Veritable core;
- correlating a percentage fill value registered by the graphical slider mechanism to a necessary confidence threshold to reach the requested percentage fill; and
- populating null values of the table at the graphical perceptible GUI until the percentage fill value is reached by selecting cell elements for which the corresponding distribution has a confidence in excess of the confidence threshold.
7. The method of claim 6, further comprising:
- receiving a 100% fill value request from the graphical slider mechanism;
- populating all null values of the table at the graphical perceptible GUI by degrading required confidence until a predicted result is available for every null value of the table.
Type: Application
Filed: Aug 29, 2013
Publication Date: Sep 18, 2014
Applicant: SALESFORCE.COM, INC. (San Francisco, CA)
Inventors: Beau David Cronin (Oakland, CA), Cap Christian Petschulat (San Francisco, CA), Eric Michael Jonas (San Francisco, CA), Fritz Obermeyer (San Francisco, CA)
Application Number: 14/014,204
International Classification: G06F 17/30 (20060101);