System, Method and Computer program Product for Accessing Hierarchical Nodes in Information Associated with Computer System Infrastructure and Assets
A method of retrieving hierarchical information, the method including receiving a request on an interface of a server, the request identifying a base node and an expansion node, the base node and the expansion node being hierarchically-arranged data objects in a data memory in communication with the server, parsing the request to determine the base node in the hierarchical information, parsing the request to determine the expansion node in the hierarchical information, retrieving a base record associated with the base node, retrieving an expansion record associated with the expansion node and referenced by the base record, inserting information from the base record and information from the expansion record into a response, and returning the response to a client computer in response to the request.
The present application claims the benefit of U.S. Provisional Patent Application No. 61/919,487, filed Dec. 20, 2013, and entitled “System, Method, and Computer Program Product for Monitoring Infrastructure and Assets,” the disclosure of which is incorporated by reference herein in its entirety.
TECHNICAL FIELDThe present disclosure relates generally to computing system monitoring and, more particularly, to performance sampling in computing systems.
BACKGROUNDInformation storage systems may include a variety of different hardware and software components. For instance, a storage system may include one or more storage controllers, where each of the storage controllers provides the low-level control for a plurality of physical storage drives. The storage system may also include network connections and other items that are ancillary to the storage functionality of the system. Storage systems continue to become more and more complex, with storage controllers hosting an increasing number of logical storage volumes and storage controllers being clustered rather than simply standing alone. There is currently a need for a management application that monitors assets of storage systems in an efficient and intuitive manner.
In the following description, specific details are set forth describing some embodiments consistent with the present disclosure. It will be apparent, however, to one skilled in the art that some embodiments may be practiced without some or all of these specific details. The specific embodiments disclosed herein are meant to be illustrative but not limiting. One skilled in the art may realize other elements that, although not specifically described here, are within the scope and the spirit of this disclosure. In addition, to avoid unnecessary repetition, one or more features shown and described in association with one embodiment may be incorporated into other embodiments unless specifically described otherwise or if the one or more features would make an embodiment non-functional.
Various embodiments of the present disclosure provide monitoring of a computer system that is both efficient and easy to understand for a human user. One embodiment includes a user interface that provides graphing of performance metric data for multiple system assets. For example, a storage system may include multiple storage drives, virtual volumes, network connections, switches, and virtual machines, among other assets. For a performance metric, such as latency, the data for that metric for multiple assets is overlaid on a graph. A human user has a convenient visual comparison tool in the overlaid graph.
Additional features of the graphing user interface may include a correlation algorithm that compares performance metric data for other assets in the computer system and selects ones of those assets with the highest correlation. The selected assets are then automatically listed for the user with an indication of correlation value. The user can also search for and select additional assets and add performance data to the overlaid data in the graphs for those selected assets as well.
To display the performance metric data, determine correlations, and to perform other tasks, the monitoring system accesses and displays information associated with storage systems and the components, assets, and elements in those storage systems. This information is often arranged hierarchically and contains descriptions of the properties that each of the assets has as well as the associations and interrelationships between and among the assets. This hierarchy of information is typically collected and stored so that it may be retrieved for later analysis and use. To support access to this information, a flexible interface for retrieving portions of the hierarchical information has been developed. The monitoring system may retrieve this hierarchical information by making one or more requests for information using the flexible interface.
The flexible interface allows the monitoring system, or other systems, to retrieve as little or as much of the hierarchical information as it desires to display particular performance metric data, screens, and/or reports. To support this flexible retrieval, information associated with each asset is kept in a record. The record contains two sections, a first section includes the properties of the asset and a second section includes references or links to other assets associated with the component. A basic retrieval request for a record would result in a response that includes both the properties and their values and the references to the other assets. Because particular display screens and reports often include some of the information from the referenced records as well, a more complex request may be made that requests not only the properties and their values, but may also ask for the information in the records of one or more of the references. This allows the monitoring system to easily retrieve information from the records of two or more associated assets with the same request. The interface also supports the ability to make more complex use of the references. Because each record in the hierarchical information typically includes references to other records associated with other assets, the interface supports requests that may specify records that correspond to records referenced by the records referenced by the base record requested, and so forth. As long as the monitoring system knows the relationships among the records, it may make an information request that includes requests for records through any number of chained-together reference linkages from the base asset. This allows the monitoring system to retrieve as little or as much of the hierarchical information describing the storage system as it desires to generate a screen or report without having to make an excessive number of requests or having to sift through large amounts of retrieved information that will not be used for the screen or report.
The management of software and firmware updates, more colloquially referred to as patches, presents significant challenges to the manager or administrator of storage and other systems. Many vendors of assets used in a storage system, such as switches, routers, storage controllers, cache memory systems, storage devices, and/or the like provide patches for updating the various assets. These patches may include fixes for errors, add new features, and so forth to the corresponding assets. Unfortunately, applying these patches does not come without its risk. Each asset receiving the patch may be configured differently so that the patch affects each asset differently. In some cases the patch may improve the functionality and/or performance of the asset and the storage system, and in other cases the patch may reduce the functionality and/or performance of the asset and the storage system. Managing and keeping track of the positive and negative impacts of each patch may become a significant burden to the storage system administrator due to the large numbers of assets in the storage system and large numbers of patches available for those assets.
The monitoring system simplifies many of the management tasks associated with patches and other updates. The monitoring system not only helps the storage system administrator apply the patch, but also keeps a record of each patch and tracks how the patch has affected the status of each of the assets the patch has been applied to. This includes determining the effects that patch has had on each asset including whether the patch has affected the ability of the monitoring system to communicate with or poll the asset and to configure the asset, as well as to determine whether the patch has had an impact on the performance of the asset. The monitoring system does this through a series of easy to use interface screens. A first interface screen facilitates application of a patch by displaying patch information to the screen including information on the types of assets to which the patch may be applied. Based on input from the storage system administrator, the monitoring system may then be used to apply the patch. After the patch is applied, the monitoring system then uses its record of the patches and the tracking of the assets to display a patch management screen that lists each patch, the number of assets that are affected, as well as summaries of any changes in status among the affected assets, and most importantly provides a recommendation on whether the patch may be approved, rolled back, or replaced by another patch. The storage system administrator may also select to see more information on any of the patches using a third screen that lists each of the affected assets, how the tracked status of the asset has changed, if at all, and makes a summary of how the patch has affected each of the assets.
Thus, by using the patch management subsystem of the monitoring system, a storage system administrator is able to quickly and easily see which patches have been applied, which assets are affected, and receive meaningful recommendations regarding whether the patches are to be kept, removed, or replaced.
The example of
The system 190 includes server system 110 connected to client system 160 via a network 165. The server system 110 accesses storage subsystems 100 that are connected to the server system 110 via a network 167. The storage subsystems 100 are included in a cluster 135. Each storage system 100 in the cluster 135 includes a set of storage devices 130 for storing client data, the storage devices 130 of the cluster 135 providing the shared storage of the storage system 100. Each storage subsystem 100 also includes a storage controller 101. Each storage controller 101 exercises low-level control over physical storage devices 130 to provide virtualized storage to server system 110 and client 160. Examples of storage hardware that can be used as physical storage devices 130 includes, e.g., hard disk drives and solid state drives, though the scope of embodiments is not limited to any particular storage hardware.
Each storage device 130 may store data from logical storage entities such as one or more storage volumes, where each volume has a file system implemented on the volume. A file system implemented on the logical storage entity may provide multiple directories in a single volume, each directory containing various filenames each of which may be mapped to a multitude of storage devices 130.
Client system 160 may run one or more applications (e.g., word processing or database programs, typified by application 161) that utilize the storage system. Client system 160 includes a computer system that interacts with server system 110 for submitting read/write access requests and for receiving or transmitting data from or to the server system 110 over the network 165. In a virtual server environment, application 161 on client system 160 may interact over the network 165 with one or more virtual machines (VMs) 115 executing on server system 110.
As mentioned above, various embodiments include a system monitoring tool that receives data from the system assets, monitors performance of the system assets, and provides user access to analyzed system data. System 190 includes a system monitoring tool that is implemented as an application. For instance, a system monitoring tool can be implemented as application 161 at client 160. Additionally or alternatively, the system monitoring tool may be implemented as one of applications 112, 117. For the purposes of this example, application 117 is described as the system monitoring tool. The system monitoring tool 117 receives system data by communicating with storage operating systems at each storage controller 101. For instance, system monitoring tool 117 may communicate via API to receive system information, such as hardware names, volume names, usage data, read and write operations per second, and the like. Various types of system information are described in more detail below. In short, the system information of this example includes any type of information that allows the monitoring tool 117 to construct a comprehensive description of the architecture and performance of system 190.
Server system 110 includes a computer system that executes applications and interacts with the client system 160 for receiving read/write access requests and receiving or transmitting data from or to the client system 160 over the network 165. Server system 110 in this example is connected to the client system 160 over a network 165 such as a local area network (LAN), an Ethernet subnet, a PCI or PCIe subnet, a switched PCIe subnet, a wide area network (WAN), a metropolitan area network (MAN), the Internet, or the like.
The server 110 may include any appropriate computer hardware and software. In one example, server 110 includes a general-purpose computer configured to execute any of a variety of operating systems, including the Unix™, Linux™, and Microsoft Windows™ operating systems.
Server system 110 includes hypervisor 113, which creates and manages one or more Virtual Machines (VMs)—in this case, VM 115. The present example shows only a single VM 115, though in other embodiments, the server 110 includes multiple VMs (not shown), each VM being used by and connected with a client 160 through computer network 165. Thus, systems with more than one client 160 may include more than one VM 115, each client being supported by at least one VM. VM 115 includes an encapsulation or instance of an operating system and applications 112 and 117 executing on top of that instance. Briefly, application 112 provides read/write access to the clients 160 to data stored in cluster 135. Application 117 is a system monitoring tool described in more detail below. In some embodiments, different types of VM hypervisors 113 may be used (e.g., VMware™ ESX, Microsoft™ Hyper-V, etc.).
Each storage system 100 is configured to allow server 110 to access its data, for example, to read or write data to the storage system. The server 110 executes application 112 that “connects” to storage systems 100 over computer network 167 to send an access request (read or write request) to storage system 100 for accessing particular data stored on the storage system 100. The VM application 112 executing on the server 110 services the connected client 160 by receiving the client access requests and submitting the access requests to the storage system 100 for execution.
The scope of embodiments is not limited to the particular architecture of system 190. For instance, other systems may include additional servers, each server being similar to server 110. While the example of
System monitoring tool 117 monitors the assets of system 190, where the assets include any hardware or software component that is included in the architecture of system 190 or affects the performance of the system 190. Examples of assets include the underlying storage drives (e.g., HDDs and SSDs), virtual volumes, storage controllers, storage subsystems, aggregates of storage subsystems, network connections, virtual machines, hypervisors, applications, and the like.
A human user interacts with system monitoring tool 117 via UI 118. UI 118 may include a command line interface, a graphical user interface (GUI), or other appropriate interface. The human user may rely on UI 118 for troubleshooting and viewing performance data. For instance, the human user may input information identifying requested performance statistics, identify new assets, and change settings using UI 118.
Storage Operating Systems (OSs) 136 run on storage controllers 100 (
System monitoring tool 117 automatically imports information on the various infrastructure assets in system 190, providing accurate and real-time visibility of servers, virtual servers, Host Bus Adaptors (HBAs), switches, storage arrays, and the like. In one example, system monitoring tool 117 discovers the assets by polling each of the assets that it is aware of. Each of the deployed assets provides one or more Application Programming Interfaces (APIs) that can be used to request information therefrom. System monitoring tool 117 is programmed to use those APIs to automatically import the information. Imported information can include, but is not limited to, device type, latency, operations per second, faults, and the like. The scope of embodiments is not limited to any particular asset information, and any appropriate asset information may be imported in various embodiments.
Display 3000 includes first graph 3001 and second graph 3002. In this example, first graph 3001 includes latency (in msec) plotted on the y-axis against time on the x-axis. First graph 3001 includes four lines, each corresponding to one of the resources with a checkmark in resource list 3004. In some embodiments, the lines in a single graph (such as the four lines in graph 3001) may be provided with a contrasting appearance, such as color coding or different types of lines, so that human user may visually discern one line from another. It is noted in graph 3001 that the four lines are overlaid within the same graph, thereby providing a human user with a convenient way to compare one resource to another.
Further in this example, second graph 3002 includes Input/Output Operations per second (IOPS) on the y-axis against time on the x-axis. Once again, there are four lines overlaid in the graph, allowing a human user to visually compare the performance of the various resources.
Display 3000 provides check boxes 3003 for a human user to select performance metrics to be displayed on graphs. In this example, the user has selected latency and IOPS, and the display 3000 includes one graph 3001 for latency and another graph 3002 for TOPS, accordingly. The user may select any (or none) of latency, IOPS, throughput (e.g., in Gb/sec), CPU usage, memory usage, and IP throughput (network throughput, e.g., in Gb/sec). The scope of embodiments is not limited to any particular set of performance metrics, as those shown in
In various embodiments graphs are plotted only for relevant performance metrics for a given resource. For example, CPU utilization is generally not relevant to Virtual Machine Disks (VMDKs), so a CPU usage chart will not show performance graph for a VMDK resource, even if the VMDK resource is selected. However, relevant metrics, such as latency, may be visually displayed for the VMDK asset in another chart.
Display 3000 includes a list of resources 3004, where each of the resources corresponds to an asset in a computer system. The resource at the top of list 3004 corresponds to a selected resource of interest (also referred to in this example as a “base resource”). The resources lower in the list 3004 are automatically selected by the system as suggested, correlated resources. The suggested resources are listed underneath the base resource in the order of their correlation percentage with the base resource. By default, the suggested resources are disabled when display 3000 first appears. When the user selects one of the suggested resources to view the performance charts (e.g., by marking a box next to the resource with a check mark), system monitoring application 117 fetches data for that suggested resource and overlays data for its relevant metrics in the charts 3001 and 3002. In one example, color coding is used so that the text for a resource in list 3004 corresponds to a color of a line in graphs 3001 and 3002.
The resource suggestions provided by display 3000 are provided to assist a human user in determining underlying causes of performance increases or decreases. A given system may have hundreds of assets, the vast majority of them uncorrelated in any useful way to a given base resource. Various embodiments provide a technique to allow a human user to focus on the few resources that are most important for explaining performance of the base resource. In this example, system monitoring application 117 automatically selects resources in the system showing a high correlation to the base resource, at least with respect to the performance metrics of interests.
In the present example, the selected performance metrics are latency and IOPS. The system monitoring application 117 selects the suggested resources based on a correlation to the base resource with respect to latency and IOPS. Indicator 3005 shows that the top-most suggested resource has a 57% correlation to the latency metric of the base resource over the time period of graph 3001.
Various embodiments may use any correlation algorithm appropriate for the resources. For instance, a conventional statistical correlation formula may be used to correlate performance metric numbers over the time period of interest. However, two resources both showing zero value for a metric over a long time period may show very nearly one-hundred percent correlation, so some embodiments may eliminate such suggestions to avoid providing useless information. An example of a statistical correlation that may be used by some embodiments includes selecting resources based on their Pearson's population correlation coefficients. The population correlation coefficient ρX,Y between two random variables X and Y with expected values μX and μY and standard deviations σX and σY is defined as:
where E is the expected value operator, coy means covariance, and, torr a widely used alternative notation for the correlation coefficient.
Display 3000 also provides more in-depth correlation explanation at tool tip 3006. In this example, the user may review how the score was calculated by selecting the score link and causing tool tip 3006 to appear. Tool tip 3006 displays which metrics (e.g. TOPS and Latency) were correlated between the different resources (e.g., LUN and VM).
At action 3110, the application displays a user interface including information regarding a base resource. For instance, the user interface may include a landing page that displays a variety of information about a selected resource, such as a description of the resource, a diagram showing connections to the resource, a graph of performance data, and the like.
At action 3120, the application runs a correlation algorithm to determine whether other resources in the computer system show a significant correlation for one or more performance metrics. In one example, the application runs a correlation algorithm for at least a subset of latency, IOPS, throughput, IP throughput, CPU usage, and memory usage and examines correlation coefficients for each of the resources for each of the performance metrics. The application examines the various resources, and if a correlation coefficient for a particular resource is significant (e.g., is above a threshold), the application selects the resource as a suggested resource.
The correlation algorithm of action 3120 can examine any metric or resource in the system. For instance, correlation may be between different computer systems (same type or different types), between different resources in different computer systems (e.g., volumes in different computer systems), and the like. In one example, the virtual machine is the base resource, and the CPU usage of the virtual machine and the latency of a storage volume that is used by the virtual machine are subject to the correlation algorithm. In another example, a storage volume is the base resource, and the its latency is correlated with traffic of a switch port.
At action 3130, the application selects one or more of the other resources as suggested resources based on results of the correlation algorithm. As mentioned above, significant correlation may include a correlation coefficient being greater than a threshold, and the application selects those resources showing significant correlation. An example list of resources is shown as list 3004 in
Also, as noted above, a resource with a performance metric at zero for a period of time may correlate highly with another resource that has the same performance metric at zero. Action 3130 may include omitting such results from the selected resources.
At action 3140, the application displays the selected resources in a list with the base resource, as in list 3004 of
At action 3150, the application overlays further performance data on the graph for a resource that was selected by the human user. As an example,
In the examples of
Returning to
Various embodiments may provide advantages over conventional systems. For instance, the overlaying of performance metric data for multiple assets on a single graph (
Furthermore, using correlation algorithms to select suggested assets for viewing by the user provides useful information to human users. While the computer system may include hundreds of resources, the correlation algorithm and provision of suggestions supplies the user with a first pass at what is probably the most relevant data to explain the performance results of the base asset.
Moreover, various embodiments also allow a user to search for and add other assets to the display, including overlaying performance data on the graphs. Such feature may give a user flexibility to view any arbitrary asset against the base asset. Such feature may be especially useful for an experienced user with knowledge of the system to look for other assets that may have a bearing on the performance of some other asset but without having passed a correlation threshold.
Network 2120 may be any kind of network including a local area network (LAN), such as an Ethernet, and/or a wide area network (WAN), such as the internet. In some examples, server 2130 may be a standalone workstation, a cluster, a production server, within a virtual machine, and/or the like. Server 2130 includes a processor 2140 coupled to memory 2150. In some examples, processor 2140 may control operation and/or execution of hardware and/or software on server 2130. Although only one processor 2140 is shown, server 2130 may include multiple processors, CPUs, multi-core processors, field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), and/or the like. Memory 2150 may include one or more types of machine readable media. Some common forms of machine readable media may include floppy disk, flexible disk, hard disk, magnetic tape, any other magnetic medium, CD-ROM, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, RAM, PROM, EPROM, FLASH-EPROM, any other memory chip or cartridge, and/or any other medium from which a processor or computer is adapted to read.
Memory 2150 may be used to store an interface 2160 and hierarchical information 2170. Interface 2160 is used by clients 2111-2119 to access the hierarchical information 2170 with clients 2111-2119 being able to make requests for all, or part, of the hierarchical information 2170. Each of the requests is directed to interface 2160 where the requested hierarchical information 2170 is retrieved, and a response is returned to the requesting client 2111-2119. Numerous mechanisms for directing the requests to interface 2160 may be used, including using a parameterized and/or unparameterized uniform resource locator (URL), using an application name corresponding to interface 2160, and/or the like. The requests may also be made using protocols or methods such as application programming interface (API) calls, remote procedure calls, representational state transfer (REST) web services, simple object access protocol (SOAP) web services, and/or the like.
As discussed above and further emphasized here,
The hierarchical information 2170 may be used to describe various objects, including the properties and/or interrelationships among the objects using one more data structures. The interrelationships may typically be represented using a tree or graph with each node representing an object and each edge representing a relationship. In some examples, each of the nodes may be stored in the hierarchical information 2170 as a record. In some embodiments, the edges may be unidirectional to describe the hierarchy in a top-down style fashion or the edges may be bidirectional to describe the hierarchy in a fashion that can be navigated in any direction. The hierarchical information 2170 may be used to organize and describe systems of any complexity from the simplest to the very complex. As the complexity of the systems being modeled increases, the numbers of nodes and edges, as well as the number of properties for each node may expand rapidly and result in a tree or graph with hundreds, thousands, or even more nodes and edges. Accessing the hierarchical information 2170 may become quite challenging. Interface 2160 may use several approaches to support access to the hierarchical information 2170 by clients 2111-2119.
One approach that interface 2160 may use is to permit access to one node of the hierarchical information 2170 at a time. Each of the requests from clients 2111-2119 includes a name, URL, identifier, and/or the like of the node of interest to interface 2160. Interface 2160 then accesses the one or more data structures storing the hierarchical information 2170, finds the requested node, and prepares a response listing each of the properties of the node, including any edges or links to other nodes in the hierarchical information. This approach leaves the problem of traversing the hierarchical information 2170 to clients 2111-2119, who control how they navigate through the hierarchical information 2170 to obtain the information of interest. As more of the hierarchical information 2170 is desired, clients 2111-2119 end up making more and more requests. In some cases this may be rather inefficient as each request and response adds overhead to the processing used to make and handle each of the requests.
Another approach that interface 2160 may use is to retrieve as much of the hierarchical information 2160 as possible, based on a node included in the request. Using the name, URL, identifier, and/or like of the node included in the request, interface 2160 recursively traverses the hierarchical information 2170 and retrieves and adds to the response as much of the hierarchical information as may be reached from the included node. In some cases, this may include each of the nodes in the hierarchical information 2170. In some embodiments, when the hierarchical information 2170 is a graph, this may add additional complexity to the recursive discovery of interface 2160 to avoid endless cycles or loops. In many cases, this approach may be rather inefficient as the response for each request may include a significant amount of the hierarchical information 2170 that the requesting client 2111-2119 is not interested in. In some examples, the requesting client 2111-2119 may also use significant computing resources to parse the large responses. In some embodiments, the request may be modified to include a maximum depth to recursively traverse in the tree or graph of the hierarchical information, but this may also result in overly large responses as clients 2111-2119 may not be interested in each of the edges from a particular node. This approach is also not effective when information associated with two unrelated or distantly related nodes, or even two nodes in different hierarchies, is desired by requesting client 2111-2119.
An approach that provides more flexibility for clients 2111-2119 when they access the hierarchical information 2170 would be desirable. To better demonstrate this, several examples of flexible requests for hierarchical information are shown using some examples of hierarchical information describing storage systems. For example, this hierarchical information may correspond to the system data for system 190 that is retrieved by system monitoring tool 117.
The by-reference section 2230 includes properties that are references to other nodes or records in the hierarchical information 2200 that are associated with node 2210. These references help build the hierarchy among the nodes.
Node 2240 is organized similarly to node 2210 and includes both a by-value and a by-reference section. The by-value section includes values for the properties associated with the storage nodes of node 2210, including representative properties for “self”, “name”, and “memory” 2242. The “memory” property 2242 demonstrates that compound by-value types may be supported as the “memory” property 2242 includes sub-properties for both “value” and “unitType”. The by-reference section includes references for both “storage” and “partner” 2244, with the “partner” reference 2244 including a link to node 2250 that is a record for the partner storage node to the storage node recorded in node 2240. Both nodes 2250 and 2260 each include by-value and by-reference sections for record properties and values for the respective nodes as well as the links to other nodes that define other parts of the hierarchy depicted in the hierarchical information 2200.
As discussed above and further emphasized here,
When the interface, such as interface 2160, receives request 2310, the request interface identifies the requested node in the hierarchy and accesses the record associated with that node, which is node 2210 in the context of
As shown, some of the values included in response 2322, such as those associated with URLs, may be altered from the values included in node 2210. As an example, the “self” by-value property is altered from the “storage/1707” in node 2210 to the full URL “/server/hierarchy/storage/1707” that corresponds to the same URL included in request 2310. This altering of URLs supports the ability for the hierarchical information 2200 to be moved from location to location without having to update the internal references as the base URL for the server and interface change. Similar alterations are also shown for the “url” properties associated with the “storageNodes” and “latency” by reference entries.
The interface generates a response 2340 to request 2330. A by-value section 2342 of response 2340 includes the same by-value information as the by-value section 2322 in response 2320. Response 2340 also includes a by-reference section 2344, introduced with the “_expands” keyword, with similar reference information as by-reference section 2324 of response 2320. One difference, however, is that the entry for the “storageNodes” reference is omitted in response 2340 because the information from corresponding node 2240 is included in response 2340, so that the “storageNodes” reference entry becomes extraneous. In some embodiments, the “storageNodes” reference entry may alternatively be included in the by-reference section 2344 to make the by-reference section complete.
Response 2340 additionally includes an inserted section 2346 where the information in node 2240 is placed. This inserted section 2346 includes both a comma-separated list of the by-value properties and values of node 2240 as well as a comma-separated list of each of the references in the by-reference section of node 2240, including the URLs for each of the referenced nodes. Thus, as
The interface may also be used to retrieve information from nodes that are associated with a chain of reference links from the base node of the request. As shown in
The interface is also able to handle additional variations in the request URL. In some embodiments, the request URL may request that multiple nodes be included in the expanded section of the result by including a comma-separated list of nodes. For example, a request with a an included URL with a parameter list of “expands=storageNodes,storageNodes.partner” would generate a response with both the expanded section 2346 and the expanded section 2664. In some embodiments, the request URL may use the dot notation to traverse a chain of references of any length. For example, “storageNodes.partner.storage” would refer to the node referenced by the storage reference in node 2250. In some embodiments, the request URL may specify a node that is not related to the base node. In some examples, the additional node may be distantly linked to the base node, unlinked to the base node, and/or even in a hierarchy different from the hierarchy of the base node.
The ability to include references to unrelated nodes, chained nodes, and multiple nodes in the “expands” parameter of the request URL provides significant flexibility in the retrieval of information from a hierarchy. This allows a client or other system the ability to request just the subset of information it desires from the hierarchy using just one request. This may reduce computing resources associated with retrieving, transmitting, and/or parsing extra requests or requests with information that is not of interest.
At a process 2410, a request is received. The request may be received from a client, such as any of the clients 2111-2119, or a system, such as system monitoring tool 117. The request may be received at an interface, such as interface 2160 using any suitable protocol or method, such as via an API call, a remote procedure call, a REST web services request, a SOAP web services request, and/or the like. The request may include a URL or other parameters and/or mechanisms that identify a base node and any reference nodes that are to be expanded in the response to the request.
At a process 2420, the base node is determined. The request is examined to determine the base node for which information is being requested. When the request is specified by an included URL, the URL may be parsed to identify the base node. In the examples of requests 2310, 2330, and 2350, the base node is the trailing part of the URL, before any parameter list, as identified by the “storage/1707” portion of the request URLs. This identifies the base node as node 2210.
At a process 2430, the base node is retrieved. Using the base node determined during process 2420, the data structure, files, databases, and/or the like containing the hierarchy of information is accessed and the record corresponding to the base node is retrieved.
At a process 2440, each of the by-value properties in the base node are iterated over and added to the response. The record retrieved during process 2430 is examined and the name and value of each of the properties in the by-value section of the record are added to the response. When the response is a string response similar to responses 2320, 2340, and/or 2360, the names and values are serialized in string form, added to a comma-separated list, and offset from the rest of the response using other delimiters such as parentheses, brackets, or curly braces. In some examples, when the value for one of the by-value properties is a compound value, such as the “memory” property 2242 of node 2240, the value portion may be offset by additional delimiters.
At a process 2450, it is determined whether the request includes a list of one or more additional nodes to expand. To support the flexible retrieval of hierarchical information, the request may also include a list of one or more nodes that are also to be included in the response. When the request includes a URL, the URL may be parsed to determine whether there is a parameter list that designates that nodes are to be expanded. In the examples of requests 2330 and 2350, a parameter list with nodes to expand is present in the URL when the parsing detects the question mark separator and the keyword “expands=”. The list of nodes to expand follows the keyword “expands=”. When the list includes more than one node, they may be separated using a comma or other separator. When the request includes nodes to expand, the nodes are expanded using a process 2460. When the request does not include nodes to expand, the base node is further processed using a process 2470.
At the process 2460, each of the nodes in the expansion list is iterated over, the corresponding node is retrieved, and the node is added to the response. The list of nodes identified during process 2450 is iterated over. For each of the nodes in the list of nodes, the corresponding node is retrieved using a process similar to process 2430, the by-value properties for the node are added to the response using a process similar to process 2440, and the by-reference properties are added to the response using a process similar to process 2470. In the examples of responses 2340 and 2360, the sections 2336 and 2444, respectively, correspond to sections of the response that may be added by process 2460. Each of the nodes in the expansion list may correspond to any node in any hierarchy that is accessible to the interface. When more than one reference or link are specified, the links may be chained together using dot notation, like the dot notation used in request 2350. After each of the nodes in the list is added to the response, the by-reference properties are added to the response using process 2470.
At the process 2470, each of the by-reference properties of the base node are iterated over and added to the response. Process 2470 may begin by adding a keyword or other separator in the response to indicate that the response now includes references that are expandable. In the examples of responses 2320, 2340, and 2360, the keyword “_expands” is used to indicate the transition to by-reference properties. The record retrieved during process 2430 is examined and the name and link for each of the references in the by-reference section of the record are added to the response. When the response is a string response similar to responses 2320, 2340, and/or 2360, the names and links are serialized in string form, added to a comma-separated list, and offset from the rest of the response using other delimiters such as parentheses, brackets, or curly braces. In some embodiments, when any of the references correspond to a node that is included in the expansion list and is already included in the response, the name and link for the corresponding reference may be omitted from the response.
At a process 2480, the response is returned. The response is returned to the client or system that made the request received during process 2410. When the request was made using an API call, the response may be included as the return value to the call. When the request was made using a remote procedure call, web service, and/or the like, the response may be returned in a response message to the client or system.
At a process 2510, a hierarchical node is selected. The preparation of documentation for a collection of records associated with hierarchical information begins when a node in the hierarchy is selected. In some embodiments, the hierarchical node may be selected by iterating through each of the hierarchical records that form one or more hierarchies. In some embodiments, the hierarchical node is selected as the node in the hierarchical information that is the head node for a tree or graph that represents the hierarchical information. In some embodiments, the hierarchical node may be selected by receiving the hierarchical node as a parameter in an API call, a web services request, and/or the like.
At a process 2520, documentation is built for each of the by-value properties of the hierarchical node. The record associated with the hierarchical node selected during process 2510 is retrieved using a process similar to process 2430. Once the record is retrieved, each of the by-value properties in the record are iterated over and corresponding documentation is built. This may include adding the name of the by-value property to the documentation including other information associated with the by-value property. This other information may include value and/or metadata information associated with the by-value property.
At a process 2530, each of the by-reference properties of the hierarchical node are iterated over, documentation is built, and the referenced node is recursively processed. The record retrieved during process 2520 is examined to determine each of the by-reference properties of the hierarchical node. Documentation is built for each of the by-reference properties that include at least a name of the by-reference property and a link, such as a web link, are added to the documentation. The link may be used to point to documentation associated with the referenced node. This documentation may be built by recursively invoking method 2500 where the referenced node becomes the hierarchical node selected during process 2510.
At a process 2540, the documentation is published. Once the documentation is assembled, it is made available to users, clients, and other systems. In some embodiments, this may be done by placing the documentation on a server where an interface may be used to access the documentation. In some examples, the documentation may be stored in a collection of files stored on a web server where the documentation for each node may be accessed and corresponding hyperlinks may be used to follow the links between nodes. In some examples, the documentation may be placed in one or more files and/or databases accessible by a help system. The help system may receive requests that identify nodes, access the files and/or databases, and retrieve the documentation associated with the requested node.
Screen 4100 may further be used to display name 4130 and description 4140 information for the patch. Screen 4100 may also provide a list of asset types 4150 to which the patch applies. The name 4130, description 4140, and/or list of asset types 4150 may be used by the user to determine whether the patch is of interest and/or to which storage system assets the patch may apply.
To facilitate application of the selected patch, screen 4100 may also include one or more controls for having the patch management tool apply the patch. In the example of screen 4100 an “Apply Patch” button 4160 is provided. When button 4160 is activated, the patch management tool may identify each the assets in the storage system of a type included in the list of asset types 4150, and then apply the selected patch to each of the identified assets. In some embodiments, the patch management tool may determine the identified assets and display them along with the list of asset types 4150 so that the user may know which specific assets may be affected by application of the patch.
Screen 4100 may also include other interface controls for managing screen 4100. For example, “Cancel” button 4170 may be used to exit screen 4100 and return to a previous interface screen.
The patch recommendation column 4220 may include one of many recommendations regarding the proposed future status of the respective patches. In some examples, the patch management tool may recommend that a patch be approved, such as is shown for the IBM SVC patch. An approval recommendation may be based on monitoring of each of the assets to which the patch has been applied to determine whether the status of each of the assets has improved or has not been adversely affected by the patch. As shown for the IBM SVC patch, application of the patch has resulted in a reduction in errors. In some examples, the patch management system may recommend that a patch be rolled back, such as is shown for the CLARION CLI patch. A roll back recommendation may be made when monitoring of the affected assets results in adverse results for the various assets. In some examples, other recommendations can include waiting for further verification of the patch, replacing the patch with a newer patch, and/or the like. In some embodiments, each of the entries in the patch recommendation column 4220 may be active interface controls that allow the user to implement the recommended action. For example, clicking on an “Approve Patch” recommendation may approve the patch and remove it from the list of monitored patches. In some embodiments, each of the entries in the patch recommendation column 4220 may include a drop-down or other menu control allowing the user to select any of the patch management actions including approve, rollback, replace, and/or the like.
As the patch management system monitors the assets to which patches are applied, it generally tracks three types of status information. Polling status determines whether the storage asset management system is able to communicate with the respective assets. In some examples, this may include the ability of the asset to respond to ping or other echo-type messages sent to the asset by the storage management system. Configuration status determines whether the storage asset management system is able to send configuration and/or provisioning instructions to the asset and have the asset be able to confirm that the configuration and/or provisioning is applied successfully. Because configuration and/or provisioning is often more complex than polling, the configuration status may reflect different types of problems and/or errors associated with different aspects of the configuration and/or provisioning. Performance status is based on monitoring of various performance metrics for the asset including latency, IOPS, throughput, CPU usage, memory usage, IP throughput, and/or the like. As with configuration status, the performance status may reflect different types of performance failures. For example, a patch may improve latency for an asset, but result in a reduction in throughput.
The entries in the conclusion column 4320 provide a summary of the differences between the pre-patch status and the post-patch status. This summary may include whether the overall status of the asset has improved (e.g., previously couldn't be polled, but is now able to be polled) or whether the status has changed (e.g., configuration is still failing, but with different errors). The entries in the conclusions column 4320 are then aggregated to form the corresponding entries in the details 4230 and recommendation 4220 columns of the patch review screen of
At a process 4410, a patch is identified. Using one or more possible input methods, a patch to be managed is identified. In some examples, this may include a user or storage system administrator using an interface control on an interface screen, such as button 4120 on screen 4100, to select and/or identify the patch.
At a process 4420, information is retrieved for the patch. Metadata and other information associated with the patch identified in process 4410 is retrieved. In some examples, this may include reading a file associated with the patch to determine the patch information. In some examples, one or more data structures, databases, and/or the like may be queried to determine the patch information. The patch information may include a name of the patch, a description of the patch, a list of asset types to which the patch may be applied, and/or the like. In some examples, the patch information may additionally include a list of assets to which the patch may be applied.
At a process 4430, the patch information is displayed. Using an interface screen, such as interface screen 4100 the patch information retrieved during process 4420 is displayed to the user.
At a process 4440, it is determined whether the patch is to be applied. The user may review the patch information displayed during process 4430 and make a determination as to whether the patch is to be applied. This decision may be based on displayed patch information and/or additional information that the user may obtain from other sources. The user may indicate an affirmative decision to apply the patch by activating a user interface control for that purpose, such as the “Apply Patch” button 4160. When the patch is to be applied, it is applied using a process 4450. When the patch is not to be applied, process 4450 may be skipped and another patch may be identified using process 4410.
At the process 4450, the patch is applied. When the patch is to be applied, the patch management tool may identify each of the assets in the storage system of a type included in the list of asset types associated with the patch that were retrieved during process 4420. This may include accessing one or more data structures, files, and/or data bases describing each of the assets in the storage system and comparing the types of those assets to the type in the list of asset types. When assets are identified with a matching asset type, the patch is applied to that asset. The patch management tool may apply the patch by sending one or more messages and/or instructions to the asset along with the patch that direct the asset to apply the patch. In some examples, as the patch is applied to each asset, the patch management tool may record this in one or more data structures, files, databases, and/or the like. Once the patch is applied to each of the identified assets, another patch may be identified using process 4410.
At a process 4510, a patch is identified. In some embodiments, one or more possible input methods may be used to identify a patch that is to be monitored. In some examples, this may include a user or storage system administrator using an interface control on an interface screen, such as a button similar to button 4120 on screen 4100, to select and/or identify the patch. In some embodiments, the identified patch may be selected from a list of patches maintained by the patch management tool in one or more data structures, files, databases, and/or the like. In some examples, the list of patches may include patches that have been applied, but have not yet been approved, rolled back, and/or replaced.
At a process 4520, the assets to which the patch is applied are determined. In some embodiments, as the patch identified during process 4510 is applied to storage assets, such as during process 4450, the patch management tool may retain a record of each of those assets and associate them with the identified patch. In some embodiments, the assets may be determined by querying the assets being managed by the storage asset management system to see whether those assets have applied the patch and/or retrieving the information from one or more data structures, files, databases, and/or the like.
At a process 4530, each of the assets to which the patch is applied is further monitored. Several sub-processes 4532-4536 are then applied to each of the assets in turn.
At the sub-process 4532, the status of the asset prior to the application of the patch is retrieved. The patch management tool accesses the one or more data structures, files, databases, and/or the like into which the storage asset management system logs status information for the assets. This includes retrieving information on whether the asset was responding to polling requests, was successfully configured, and/or was demonstrating suitable performance during a time period prior to the application of the patch. The retrieved information may further include information about different types of errors received when monitoring and/or managing the asset and/or performance data associated with the asset.
At the sub-process 4534, the status of the asset after the application of the patch is retrieved. Similar to sub-process 4532, status information related to polling, configuration, and/or performance associated with the asset during a time period after the patch was applied is retrieved.
At the sub-process 4546, effectiveness of the patch is determined and summarized. The patch management tool makes one or more comparisons between the retrieved status information from both before and after when the patch was applied. Based on changes in the status, including the polling, configuration, and/or performance capabilities of the asset, the effectiveness of the patch is determined for the asset and a summary is generated. In some embodiments, the effectiveness of the patch and the summary may be sufficient to fill in a row of a patch asset review list similar to the patch asset review list 4300.
At a process 4540, a patch recommendation is made. The patch management system aggregates the patch effectiveness and summary determined during sub-process 4546 to make a recommendation regarding whether the patch is to be approved, rolled back, replaced, and/or the like. In some embodiments, the recommendation may be based on counts of how many of the assets are positively affected by the patch versus how many of the assets are negatively affected by the cache. When all and/or a majority of the assets are positively affected by the patch, the recommendation may be to approve the patch. When a majority and/or even some of the assets are negatively affected by the path, the recommendation may be to roll back and/or replace the patch. In some examples, a recommendation to replace the patch may additionally be based on whether another, potentially newer, patch is available for each of the assets to which the patch is applied. In some examples, when insufficient information is available to determine asset status after application of the patch, the recommendation may include waiting for further status monitoring. In some embodiments, the recommendation and/or aggregation may be sufficient to fill in a row of a patch review list similar to patch review list 4200.
In some embodiments, the patch management tool may further support implementation of the recommendation. For example, when the recommendation is roll back and is approved by the user, the patch management system may roll back the patch by sending one or more messages and/or instructions to the asset instructing the asset to roll back the patch.
It should be noted that the examples above are given in the context of a network storage system, through the scope of embodiments is not so limited. Rather, the concepts described above may be implemented in any type of computing cluster, wherein performance data is sampled and analyzed. One example embodiment includes a cluster of server nodes, where performance data for the server nodes themselves, as well as for the applications running on the server nodes, is sampled according to a workload of each node or application. Process 400 would transfer the sampled data to an analysis application for further processing.
When implemented via computer-executable instructions, various elements of embodiments of the present disclosure are in essence the software code defining the operations of such various elements. The executable instructions or software code may be obtained from a non-transient, tangible readable medium (e.g., a hard drive media, optical media, RAM, EPROM, EEPROM, tape media, cartridge media, flash memory, ROM, memory stick, network storage device, and/or the like). In fact, readable media can include any medium that can store information.
In the embodiments described above, example clients 160, server 110, storage controllers 101, and server 2130 include processor-based devices and may include general-purpose processors or specially-adapted processors (e.g., an Application Specific Integrated Circuit). Such processor-based devices may include or otherwise access the non-transient, tangible, machine readable media to read and execute the code. By executing the code, the one or more processors perform the actions of the processes of
Although illustrative embodiments have been shown and described, a wide range of modification, change and substitution is contemplated in the foregoing disclosure and in some instances, some features of the embodiments may be employed without a corresponding use of other features. One of ordinary skill in the art would recognize many variations, alternatives, and modifications. Thus, the scope of the invention should be limited only by the following claims, and it is appropriate that the claims be construed broadly and in a manner consistent with the scope of the embodiments disclosed herein.
Claims
1. A method of retrieving hierarchical information in response to a request from a client computer, the method comprising:
- receiving the request on an interface of a server, the request identifying a base node and an expansion node, the base node and the expansion node being hierarchically-arranged data objects in a data memory in communication with the server, the base node and the expansion node representing assets of a computer system being monitored;
- parsing the request to determine the base node in the hierarchical information;
- parsing the request to determine the expansion node in the hierarchical information;
- retrieving a base record associated with the base node;
- retrieving an expansion record associated with the expansion node and referenced by the base record;
- inserting information from the base record and information from the expansion record into a response; and
- returning the response to a client computer after inserting the information from the base record from the expansion record.
2. The method of claim 1 wherein inserting the information associated with the base record into the response comprises:
- inserting one or more by-value properties of the base record into the response; and
- inserting one or more by-reference properties of the base record into the response.
3. The method of claim 2 wherein inserting the by-value properties into the response comprises inserting a name and value for each of the by-value properties.
4. The method of claim 2 wherein inserting the by-reference properties into the response comprises inserting a name and a reference to another node of the hierarchy for each of the by-reference properties.
5. The method of claim 2 wherein the by-value properties are separated from the by-reference properties in the response by an expansion keyword.
6. The method of claim 1 wherein determining the base node comprises parsing a uniform resource location (URL) included in the request to determine an identifier associated with the base node.
7. The method of claim 1 wherein determining the expansion node comprises parsing a uniform resource locator (URL) included in the request to detect an expansion parameter, the expansion parameter including information associated with an identifier of the expansion node.
8. The method of claim 1 wherein when the expansion node corresponds to a reference included in a by-reference property of the base record.
9. The method of claim 1 wherein the expansion node is directly linked to the base node in the hierarchy by a uniform resource locator (URL).
10. The method of claim 1 wherein the expansion node is indirectly linked to the base node in the hierarchy via a chain of references from the base node.
11. A server computer in a network, the server computer comprising:
- a processor; and
- memory accessible by the processor and storing computer-readable instructions, the processor performing the following actions by executing the instructions: receiving a request for hierarchical information from a client, the hierarchical information including a base node and an expansion node linked to the base node, further wherein the request identifies the base node and the expansion node; retrieving a base record associated with the base node in response to the request, the base record including a link to an expansion record associated with the expansion node; retrieving the expansion record in response to the request; formulating a response to the request by inserting information from the base record into the response and inserting information from the expansion record into the response, where the information from the expansion record includes references to other nodes in the hierarchical information; and returning the response to the client.
12. The server of claim 11, wherein the server is in communication with the client by a network.
13. A computer program product having a computer readable medium tangibly recording computer program logic for retrieving hierarchical information associated with assets of a network storage system, the computer program product comprising:
- code to receive a request for portions of the hierarchical information, the request identifying a first node and a second node, the first node being higher in the hierarchical information than the second node by referencing the second node as a property of the first node;
- code to retrieve a first record associated with the first node in response to the request, where the first record includes a link to the second record;
- code to retrieve the second record;
- code to insert first record information into a response to the request, where the first record information includes by-reference properties of the first record and by-value properties of the first record;
- code to insert second record information into the response; and
- code to return the response to the client.
14. The computer program product of claim 13 wherein the by-value properties include a name and value for each of the by-value properties.
15. The computer program product of claim 13 wherein the by-reference properties include a name and a reference to another node of the hierarchy for each of the by-reference properties.
16. The computer program product of claim 13 wherein the by-value properties are separated from the by-reference properties in the response by an expansion keyword.
17. The computer program product of claim 13 wherein the code to retrieve the first record comprises:
- code to parse a uniform resource location (URL) included in the request to determine an identifier associated with the first node.
18. The computer program product of claim 13 wherein code to retrieve the second record comprises:
- code to parse a uniform resource locator (URL) included in the request to detect an expansion parameter, the expansion parameter including information associated with an identifier of the second node.
19. The computer program product of claim 13 wherein the second node is directly linked to the first node in the hierarchy by a uniform resource locator (URL).
20. The computer program product of claim 13 wherein the second node is indirectly linked to the first node in the hierarchy via a chain of references from the first node.
Type: Application
Filed: Jun 20, 2014
Publication Date: Jun 25, 2015
Inventors: Ophir Horn (Waltham, MA), Ran Gilboa (Waltham, MA), Jason Sprague (Waltham, MA), Dmitry Erkin (Waltham, MA), Francisco Rosa (Waltham, MA)
Application Number: 14/311,011