ENTERPRISE SEARCH METHOD AND SYSTEM
A system and method for enterprise search includes one or more computer-readable media storing computer-executable instructions that, when executed on one or more processors that perform acts including extracting one or more of term data, personal data and metadata from one or more predetermined resources; retrieving a set of information derived from the extracted term data, personal data and metadata responsive to a query; and receiving feedback responsive to the set of information, the feedback augmenting at least one of the one or more predetermined resources.
The discussion below is merely provided for general background information and is not intended to be used as an aid in determining the scope of the claimed subject matter.
A large amount of electronic documents are prevalent today throughout organizations and on the Internet. These documents contain useful informational elements for a number of different purposes. In addition to explicit text within the documents, the documents include metadata that is indicative of a particular information retrieval element in a document. There are many different types of metadata for a document including title, author, date of creation, etc. Metadata in documents is useful for many kinds of document processing applications including search, browsing and filtering. Metadata can be defined by an author of the document to be used in these processing applications. However, authors seldom define document metadata appropriately. A search using only metadata, therefore provides limited results.
Other types of documents are available in organizations for intranet search to research people in an organization, and to research technical data. Documents, homepages and the like provide resources for search queries. Current methods of searching different types of data do not provide efficient search results appropriate for the type of search performed. In an Enterprise level search access to information for users is exacerbated. For example, most of the current enterprise systems try to solve this problem by using relevance search and display the results as a list of URLs. What is needed is a system and method that accounts for the type of search requested.
SUMMARYThis Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
Embodiments herein are directed to methods and systems that relate to extracting data from resources to respond to a query. The extracted data can include frequently asked question (FAQ) data, term data, personal data and metadata. The methods and systems also relate to retrieving a set of information derived from the extracted FAQ data, term data, personal data and metadata responsive to a query, and receiving feedback responsive to the set of information. In an embodiment, the feedback may augment at least one of the one or more predetermined resources.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter. The term “tools,” for instance, may refer to system(s), method(s), computer-readable instructions, and/or technique(s) as permitted by the context above and throughout the document.
The detailed description is described with reference to accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different figures indicates similar or identical items.
The subject matter discussed below relates generally to processing electronic documents. In one aspect, features are identified from a set of training documents to aid in extracting information from documents to be processed. The features relate to formatting information and linguistic information in the document. One or more models are developed to express these features. During information extraction, documents are processed and analyzed based on at least one of the models and information is extracted based on the models. The invention is operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the invention include, but are not limited to, personal computers, server computers, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
The invention may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices. Tasks performed by the programs and modules are described below and with the aid of figures. Those skilled in the art can implement the description and figures as processor executable instructions, which can be written on any form of a computer readable medium.
Term information extraction module 112 may include an extraction of definitions and acronyms from documents. In an embodiment, the definition and acronym extraction may occur in three steps. In the first step, texts are extracted from WORD and HTML documents. Webpage documents can include the extensions: .ascx; .aspx; .asp; .htm; .html. Office documents can include the extensions .doc; .docm; .docx; .dot; .ppt; .pptm; .pptx; .pps and .txt file.
In the second step, definitions and acronyms may be extracted. Terms, including base noun phrases, in definitions are recognized in as long as they include expansions of acronyms. In the third step, all extracted information may be saved into term resource database 114.
Term resource database 114 may include a database tables of definitions with related terms and acronyms with related expansion. In an embodiment, each definition can only have one related term. Likewise, each acronym expansion may only have one acronym.
Term information extraction module 112 may be coupled to term resource database 114, which is available to offline extraction module 102. Another source of data available to offline extraction module 102 includes documents available via an enterprise Intranet 116.
Within offline extraction module 102, operation modules may include metadata extraction module 120, expert and expertise extraction module 122 and definition acronym, home page, and frequently asked questions (FAQ) extraction module 124.
Referring to Table 1, below, a data schema for data extracted is illustrated:
Offline extraction module 102 is coupled to knowledge database 130 for storing organized extracted data in different formats. One format stored in knowledge database 130 includes user feedback from user feedback database 132 that is shown coupled to term resource database 114, which is coupled to offline extraction module 102.
Knowledge database 130 supplies organized search results to online search module 140, which includes organized search module 142 and wiki search module 146. Organized search module 142 responds to enterprise search user interface 150 by providing organized search results. Wiki search module 146 provides a search option for users who wish to leverage their search with prior user feedback.
FAQ extraction module 440, in one embodiment, operates uses a two-stage model to extract question and answer pairs from documents of Word. First, candidate FAQ documents are collected by locating “faq” in a URL, “faq” in a title, “Frequently Asked Questions” in a title, “help” in a URL, and “help” in a title.
If a file, whose name ends with “.doc” or “.docx”, meets any of the FAQ requirements, it is a candidate FAQ document. All candidate documents' paths are listed in a file.
Next, question/answer pairs are extracted after text files are created from the documents. In the extracted text file, each doc starts with <message> and ends with </message>. File path of each doc is between <link> and </link>. Content of each doc is between <text> and </text>.
Each text is divided into lines and all operations are conducted on the lines. In the code, function ReadUnits( ) is used to read text line by line. That returns a Boolean value “true” if success, otherwise false.
Four rules are used to count the score of each line: (1) bool Rule1(int LineNumber). If a line starts with a question word (such as how, what etc), this function will return true, otherwise false.
(2) bool Rule2(int LineNumber) If a line ends with question mark, this function will return true, otherwise false.
(3) bool Rule3(int LineNumber) If a line starts with a pattern like “Q” or “Q.” or “Q:”, function returns true, otherwise false.
(4) bool Rule4(int LineNumber) If a line starts with a number (such as “4.”, “11.” etc), function returns true, otherwise false.
The score of a line is counted by define an integer s=0; If this line meets Rule 1, s=s+1; If this line meets Rule 2, s=s+2; If this line meets Rule 3, s=s+1; If this line meets Rule 4, s=s+1; If the number of characters in this line is less than 15 or greater than 200, s=0; For each line in the text, we calculate a score for it. If one line gets a score not less than 2 it will be added into question list. At last, if some consecutive lines are questions, they will be removed from question list.
To produce answers, lines between questions are assumed to be answers.
FAQ extraction module 440, in one embodiment, operates uses either a one or a two stage model to extract question and answer pairs from HTML pages. In a two stage model, candidate FAQ HTML pages are collected by locating “faq” in a URL, “faq” in a title, “Frequently Asked Questions” in a title, “help” in a URL, and “help” in a title.
Next, question and answer pairs are extracted by parsing each candidate page to extract the pairs.
In another embodiment, a one-stage mode is used to extract question and answer pairs by skipping the stage of locating candidate HTML pages because only a very small portion of FAQ pages can be selected through the first step. Many pages don't meet any of those rules but they contain question and answer pairs.
Tag and text based FAQ extraction module 480 outputs two lists, one for questions and the other for corresponding answers. First, module 480 preforms preprocessing to remove white space (including ‘\t’, ‘\r’, ‘\n’) not between <pre> and </pre> from original text. Next, the extraction module 480 removes strings between <script> and </script> or <style> and </style> and substitutes any “ ” with space.
Next, text is split into units by special HTML tags (including <br>, <p>, </p>, <div>, </div>, <h[1-6]>, </h[1-6]>, <td>, <dt>, <dd>, <li>). Next, to save running time, this function records text of each unit after removing tags and calculates in advance which units meet these four rules: (1) bool IsHyperlink(string str) If str is a hyperlink, this function will return true, otherwise false. (2) bool Rule1(int line) If the unit's text begins with “Q” and ends with question mark, this function will return true, otherwise false. (3) bool Rule2(int line) If the unit have both pre-tag (<strong><b><u><em><td>) and corresponding post-tag (</strong></b></u></em></td>), and its text ends with question mark, this function will return true. (4) bool Rule3(int line) If the unit has only pre-tag (<li><dt><p><h[1-6]><td><span><div><font>) and its text ends with question mark, this function will return true, otherwise false.
Next, for each unit, if it meets any of Rule1( ), Rule2( ) or Rule3( ) module 480 assumes the unit is a candidate question. The surrounding tags are extracted as candidate patterns.
Next, the extraction module 480 performs pattern validation. First, the module counts the score of each candidate unit. Next, the module divides all candidate units into groups by their patterns. Units in each group should have the same pattern. Module 480 calculates an average score of each group, and sort groups by average score.
Select candidate patterns from top N groups are determined to be in a final question pattern set. Other patterns are discarded. The number of questions with final question patterns is then set to n1.
The module then searches all units with the pattern in the final question pattern set. If its score is greater than zero, this unit will be added into the question list. Let the number of added questions be n2. If n2 is larger than 2 times of n1, all questions are discarded. In other words, if too many units have the same patterns with the units in top N groups, these patterns are too general, so no FAQs are extracted from the unit.
The score of each unit is counted by defining an integer, s=0. If this unit meets rule1, then s=s+2; If this unit meets rule2, then s=s+1; If this unit meets rule3, then s=s+1; If character amount in this unit is greater than 200, then s=s−1. If sentence number of this unit is between 3 and 6, then s=s−1. If sentence number of this unit is greater than 7, then s=s−2
As described above, any units between two questions is assumed to be an answer. According to an embodiment, all tags are removed from question units and remaining text is provided to a question list. Likewise, all tags are removed from answer units and remaining text is added to an answer list.
FAQ results are provided from text-based FAQ extraction 440 and Tag and text based FAQ extraction 480 to FAQ saving 450 and then to FAQ database 490, which can be incorporated into knowledge database 130 shown in
Referring to
Referring back to
According to an embodiment, FAQ database 490 hold FAQ formatted data for an enterprise responsive to a FAQ Search by users. The FAQ's extracted by extraction modules 440 and 480 can be from different sources such as discussion lists, technical notes, online forums, and the like.
FAQ extraction by modules 440 and 480 can be conducted by a rule based method. FAQ's can be stored in the database 490 together with other knowledge such as definition, acronym. When the users input a query, the system will return the relevant FAQ's.
FAQs are searched based on a number of factors: the matching degree between the query and the question, length of the question, number of clicks, and the like. A linear combination of the factors can be used.
Referring now to
The ranking sore is then provided to both extract expertise information module 630 and to extract expert information module 640. “Term to person relation” is equivalent to an expert search. Module 640 counts the occurrences of “person in author, term in title”, “person in author, term in body”, and “person in body, term in body”. An embodiment is directed to designating a person as an expert for a given search as a function of the number of occurrences.
In another embodiment, module 640 could perform a “Person to term relation” representing the expertise of a person. The person is viewed as one familiar with a term if the term occurs frequently in his documents.
Extracted expertise information is provided to term to person relation database 650 and extracted expert information is provided to person to term relation database 660.
Referring now to
Specifically, relationship data is determined for database 610 by first collecting data from each of databases metadata 460, homepage 702, query log 706, definition and acronym database 720, personal information resource 108 and term resource 114 in addition to text from documents 722. Homepage 702, query log 706 and definition and acronym 720 each are coupled to term extraction and merge module 112 to create term resource database 114. metadata database 460 and term resource database 114 each provide data to “Find term in document title” database 730. Term resource database 114 and text from documents 722 each provide data to “Find term in document body” module 732. Personal information 108 and text from documents 722 each provide data to “Find person in document body 734”; and metadata 460 provides data to “Find Document Author” module 736. The results from each of modules 730, 732, 734 and 736 provide the data for term, person, document relation 610.
Referring back to
In the retrieved results, users can add tags to homepages and documents. The user can also modify or delete the tags provided by him/her. Tags are used for searches of tagged documents including tagged homepages.
In one embodiment, users can add additional information to the search result, for example, document, definition, and homepage. The format of input depends on the type of search result, i.e., the type of resource related to the search result. The user can also modify his/her input data. Inputs from different users can be saved and shown separately. In the retrieved results, user can add tags to homepages and documents. Tags can be used for searches of tagged documents including tagged homepages.
In one embodiment, users can vote on their search result. For example, an add thumb up or thumb down can be included with user interface 150. This information can be aggregated and used in ranking of search results for later searches. In one embodiment, the voting can be conducted on a query and search result pair to provide a relevance of the query and search result and not the quality of the search result. Voting from a user can be modified by the same user. The voting information can be saved into a database. In an embodiment, a user can only vote on a query search result pair. The voting results from the same user are only counted once.
User feedbacks can be displayed in the user search result page of user interface 150. In one embodiment, feedback is also saved to improve the offline extraction module 102 accuracy.
In an embodiment, a user can add additional information to the search result, for example, document, definition, and homepage. The format of input depends on the type of search result, i.e., section of search result. The user can also modify his/her input data. The inputs from users can be validated at the input time. For example, on the length of homepage title.
Inputs for the same query from different users can be allowed in the system, even though there are contradictions (an embodiment assumes that the motivations of users' inputs are to improve the system). For example, a user may add “http://msw” as the home page of the query “msw”, but user B may add “http://msw” as the home page of “Microsoft”. All user inputs are saved into database online.
Inputs from different users are saved and shown separately. An input from the same user can be modified by the user. User input is not a wiki, i.e., a result is not from multiple users, but a single user. The user input data types are illustrated below in Table 2. An exemplary user feedback data schema is shown in Table 3.
Referring back to
Index should be updated when the data is changed. There are two possible strategies for index update: batch mode and online mode. Decision “query matched”? 830 separates the query so that the query is parsed; related result is retrieved; and the related queries are also suggested. If the query matches, query suggesting module 840 provides suggested queries 852. Also, if the query matches, types results retrieval module 842 provides an organized search results 854. If there is no match at block 830, relevance integration module 846 provides general search results 858.
After results are found, online search module 140 ranks retrieved results by qualities, such as a number of documents for homepages. Additionally, user voting results may also be included as a quality metric. According to an embodiment, exact matched results are ranked highly. The exact matched results, in one embodiment, may be leveraged by a full text search feature at SQL Server. Extracted results and user feedback results are aggregated and shown to users at enterprise search user interface 150.
According to an embodiment, tagged documents are included with a relevance search. An embodiment is directed to rules for ranking result including, ranking user input results are always ranked before the automatically extracted results. Results with more votes are ranked before those with fewer votes. The ranking may apply within user input results or automatically via extracted results respectively. If the results have the same votes, an embodiment is directed to ranking by a default ranking function. The default ranking function depends on each resource.
Default ranking functions for a homepage resource include prioritizing by determining whether the homepage is automatically extracted or by user input. (user input takes the higher priority). Second, user rank (voting) is determined. Third, the ranking takes into account the creation time (from new to old). Last, the ranking takes into account the length of Homepage's title. (Ascending order).
Default ranking for definition resources first determines whether the Definition is system extracted or by user input. (user input takes the higher priority). Second, the ranking takes into account a user rank (voting). Third, the creation time is determined. Last an auto rank by system extraction score is taken into account.
Default ranking for acronym resources first determines whether the acronym is system extracted or by user input (user input takes the higher priority). Second, the ranking takes into account a user rank (voting). Third, the creation time is determined. Last an auto rank by system extraction score is taken into account.
Default ranking for Expertise resources first determines whether the expertise data is system extracted or by user input (user input takes the higher priority). Second, the ranking takes into account a user rank (voting). Third, the evidence count, such as a document count of a term-people relation is determined. Last a creation time is taken into account.
Default ranking for FAQ resources first determines whether the FAQ data was system extracted or by user input (user input takes the higher priority). Second, the ranking takes into account a user rank (voting). Third, a creation time is taken into account in ascending order.
Default ranking for Authored Document resources first determines whether the Authored Document data was system extracted or by user input. (user input takes the higher priority). Second, the user ranking of the person-document relation by voting is taken into account. Last, a creation time is taken into account.
Default ranking for a co-author, co-occurrence or colleague resources first determines whether the co-author, co-occurrence or colleague resource data was system extracted or by user input with user input taking the higher priority. Second, the user ranking by voting is taken into account. Third, an evidence count for a co-author or co-occurrence document count is performed for people to people relations. Last, a creation time is taken into account.
Default ranking for expertise resources first determines whether the expertise resource data was system extracted or by user input. (user input takes the higher priority). Second, the user ranking by voting is taken into account. Third, an evidence count/document count is performed for people to term relations. Last, a creation time is taken into account.
For tagged results, a tagged homepage, rank is determined by user rank via voting first. Then, a document count for the particular homepage is determined.
For tagged documents, rank is determined by user rank via voting first. Then, a creation time for the tagged document is taken into account.
Referring now to
As shown in
As shown in
Further detailed information is available by using one of the tabs identified as “more . . . ” 1050.
Referring now to
Block 1130 provides for retrieving a set of information derived from the FAQ data, term data, personal data and metadata responsive to a query. For example, online search module 140 provides for retrieving a set of information, which may be in sections appropriate to the predetermined source that are responsive to a query receive via enterprise search user interface. The set of information may include FAQs, relevance document results, organized search results and the like.
In one embodiment, the retrieving includes extracting one or more question and answer pairs from the one or more predetermined resources and organizing the one or more answer pairs in a database table of FAQ data.
In one embodiment, the retrieving includes organizing the FAQ data, the term data, personal data and the metadata using machine learning and/or data mining.
Block 1140 provides for receiving feedback responsive to the set of information, the feedback augmenting at least one of the one or more predetermined resources. For example enterprise search user interface 150 may receive feedback that is provided to online search module 140 and to user feedback database 132. From user feedback database 132, term resource database 114 may provide the feedback to offline extraction module 102 to augment knowledge database 130.
With reference to
Computer 1210 typically includes a variety of computer readable media. Computer readable media may be any available media that may be accessed by computer 1210 and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer readable media may comprise computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital video disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which may be used to store the desired information and which may be accessed by computer 1210. Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of any of the above should also be included within the scope of computer readable media.
The system memory 1230 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 1231 and random access memory (RAM) 1232. A basic input/output system 1233 (BIOS), containing the basic routines that help to transfer information between elements within computer 1210, such as during start-up, is typically stored in ROM 1231. RAM 1232 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 1220. By way of example, and not limitation,
The computer 1210 may also include other removable/non-removable volatile/nonvolatile computer storage media. By way of example only,
The drives and their associated computer storage media discussed above and illustrated in
A user may enter commands and information into the computer 1210 through input devices such as a keyboard 1262, a microphone 1263, and a pointing device 1261, such as a mouse, trackball or touch pad. Other input devices (not shown) may include a joystick, game pad, satellite dish, scanner, or the like. These and other input devices are often connected to the processing unit 1220 through a user input interface 1260 that is coupled to the system bus, but may be connected by other interface and bus structures, such as a parallel port, game port or a universal serial bus (USB). A monitor 1291 or other type of display device is also connected to the system bus 1221 via an interface, such as a video interface 1290. In addition to the monitor, computers may also include other peripheral output devices such as speakers 1297 and printer 1296, which may be connected through an output peripheral interface 1295.
The computer 1210 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 1280. The remote computer 1280 may be a personal computer, a hand-held device, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 1210. The logical connections depicted in
When used in a LAN networking environment, the computer 1210 is connected to the LAN 1271 through a network interface or adapter 1270. When used in a WAN networking environment, the computer 1210 typically includes a modem 1272 or other means for establishing communications over the WAN 1273, such as the Internet. The modem 1272, which may be internal or external, may be connected to the system bus 1221 via the user-input interface 1260, or other appropriate mechanism. In a networked environment, program modules depicted relative to the computer 1210, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.
Claims
1.-20. (canceled)
21. A method implemented by one or more computing devices, the method comprising:
- receiving a query from a user;
- searching for one or more question and answer pairs from one or more predetermined resources based at least in part on one or more matching degrees between the query and respective one or more questions of the one or more question and answer pairs; and
- providing at least respective one or more answers of the one or more question and answer pairs for presentation to the user.
22. The method of claim 21, wherein searching for the one or more question and answer pairs is further based on one or more lengths of the respective one or more questions, respective numbers of clicks of the one or more question and answer pairs, or respective numbers of votes associated with the one or more question and answer pairs.
23. The method of claim 21, further comprising receiving feedback from the user responsive to the respective one or more answers, the feedback being indicative of a quality of at least one of the respective one or more answers.
24. The method of claim 21, further comprising receiving feedback from the user responsive to the respective one or more answers, the feedback being at least one of a correction to at least one of the respective one or more answers, a vote associated with the at least one of the respective one or more answers, or an addition to the at least one of the respective one or more answers.
25. The method of claim 21, wherein the one or more predetermined resources comprise a plurality of question and answer pairs, and the method further comprises organizing the plurality of question and answer pairs using at least one of machine learning or data mining.
26. The method of claim 21, further comprising generating frequently asked question (FAQ) data.
27. The method of claim 26, wherein generating the FAQ data comprises:
- collecting candidate files based at least in part on locating one or more keywords associated with FAQs in the candidate files;
- extracting candidate patterns for the FAQs from the candidate files based at least in part on one or more rules associated with the FAQs; and
- validating the candidate patterns for the FAQs.
28. One or more memory devices storing executable instructions that, when executed by one or more processors, cause the one or more processors to perform acts comprising:
- receiving a query from a user;
- searching for one or more question and answer pairs from one or more predetermined resources based at least in part on one or more matching degrees between the query and respective one or more questions of the one or more question and answer pairs; and
- providing at least respective one or more answers of the one or more question and answer pairs for presentation to the user.
29. The one or more memory devices of claim 28, wherein searching for the one or more question and answer pairs is further based on one or more lengths of the respective one or more questions, respective numbers of clicks of the one or more question and answer pairs, or respective numbers of votes associated with the one or more question and answer pairs.
30. The one or more memory devices of claim 28, the acts further comprising receiving feedback from the user responsive to the respective one or more answers, the feedback being indicative of a quality of at least one of the respective one or more answers.
31. The one or more memory devices of claim 28, the acts further comprising receiving feedback from the user responsive to the respective one or more answers, the feedback being at least one of a correction to at least one of the respective one or more answers, a vote associated with the at least one of the respective one or more answers, or an addition to the at least one of the respective one or more answers.
32. The one or more memory devices of claim 28, wherein the one or more predetermined resources comprise a plurality of question and answer pairs, and the method further comprises organizing the plurality of question and answer pairs using at least one of machine learning or data mining.
33. The one or more memory devices of claim 28, the acts further comprising generating frequently asked question (FAQ) data.
34. The one or more memory devices of claim 33, wherein generating the FAQ data comprises:
- collecting candidate files based at least in part on locating one or more keywords associated with FAQs in the candidate files;
- extracting candidate patterns for the FAQs from the candidate files based at least in part on one or more rules associated with the FAQs; and
- validating the candidate patterns for the FAQs.
35. A system comprising: one or more processors; memory storing a plurality of modules that, when executed by the one or more processors, cause the plurality of modules to perform acts comprising:
- receiving a query from a user;
- searching for one or more question and answer pairs from one or more predetermined resources based at least in part on one or more matching degrees between the query and respective one or more questions of the one or more question and answer pairs; and
- providing at least respective one or more answers of the one or more question and answer pairs for presentation to the user.
36. The system of claim 35, wherein searching for the one or more question and answer pairs is further based on one or more lengths of the respective one or more questions, respective numbers of clicks of the one or more question and answer pairs, or respective numbers of votes associated with the one or more question and answer pairs.
37. The system of claim 35, the acts further comprising receiving feedback from the user responsive to the respective one or more answers, the feedback being indicative of a quality of at least one of the respective one or more answers.
38. The system of claim 35, the acts further comprising receiving feedback from the user responsive to the respective one or more answers, the feedback being at least one of a correction to at least one of the respective one or more answers, a vote associated with the at least one of the respective one or more answers, or an addition to the at least one of the respective one or more answers.
39. The system of claim 35, wherein the one or more predetermined resources comprise a plurality of question and answer pairs, and the method further comprises organizing the plurality of question and answer pairs using at least one of machine learning or data mining.
40. The system of claim 35, the acts further comprising generating frequently asked question (FAQ) data, the generating comprising:
- collecting candidate files based at least in part on locating one or more keywords associated with FAQs in the candidate files;
- extract candidate patterns for the FAQs from the candidate files based at least in part on one or more rules associated with the FAQs; and
- validate the candidate patterns for the FAQs.
Type: Application
Filed: May 1, 2017
Publication Date: Aug 17, 2017
Inventors: Hang Li (Beijing), Yunhua Hu (Beijing), Xin Zou (Beijing), Xiaoyuan Cui (Beijing), Guangping Gao (Beijing), Dmitriy Meyerzon (Bellevue, WA), Victor Poznanski (Samammish, WA)
Application Number: 15/583,499