SCALING DISTRIBUTED COMPUTING SYSTEM RESOURCES BASED ON LOAD AND TREND
The described technology is generally directed towards automatically scaling distributed computing resources of a distributed computing system based on a system load measurement and a trend factor indicative of whether the system load is increasing or decreasing. If a computing resource load value is above a resource addition threshold value and the trend factor indicates that the computing resource load is increasing, a corresponding computing resource is added to the distributed computing system. If a computing resource load value is below a resource removal threshold value and the trend factor indicates that the computing resource load is decreasing, a corresponding computing resource is removed from the distributed computing system. The trend factor can be obtained using a moving average convergence divergence (MACD) direction indicator.
The subject application generally relates to distributed computing systems, and, for example, to automatically scaling distributed computing system resources based on current system load and a trend factor, and related embodiments.
BACKGROUNDIn a distributed computing system, load scalability refers to the ability of the distributed computing system to handle increasing or decreasing workloads, by scaling up the resources in use to accommodate heavier loads, or scaling down the resources to accommodate lighter loads (and thereby conserve resources). A typical way in which to decide whether to add or remove resources is based on average system load utilization, e.g., evaluated against threshold values.
However, at times the system load average can be deceiving. As a result, basing scalability decisions on the system load average sometimes causes resources to be unnecessarily added, and at other times causes resources to be prematurely deleted.
SUMMARYThis Summary is provided to introduce a selection of representative concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used in any way that would limit the scope of the claimed subject matter.
Briefly, one or more aspects of the technology described herein are directed towards evaluating a computing resource load value that represents a computing resource load in a distributed computing system, and a trend factor that indicates whether the computing resource load is increasing or decreasing. In response to the computing resource load value being above a resource addition threshold value and the trend factor indicating that the computing resource load is increasing, a computing resource is added to the distributed computing system. Further, a computing resource can be removed from the distributed computing system, e.g., in response to the computing resource load value being below a resource removal threshold value and the trend factor indicating that the computing resource load is decreasing.
Other embodiments may become apparent from the following detailed description when taken in conjunction with the drawings.
The technology described herein is illustrated by way of example and not limited in the accompanying figures in which like reference numerals indicate similar elements and in which:
Various aspects described herein are generally directed towards a predictable automatic scaling technology, which based on a measure of the system load (e.g., of any resource or combination of resources) in conjunction with a trend and direction of the system load. The use of the direction and trend makes the automatic scaling decisions more intelligent in practical application usage.
By way of example, consider that the average system load is measured and reported to be above a resource addition threshold. Simply adding resource(s) in this state, as is typically done, can unnecessarily and wastefully add resources if the system load is decreasing. Similarly, if the average system load is measured and reported to be below a resource removal threshold, removing resource(s) in this state, as is typically done, is premature if the load is system increasing and the removed resource(s) are going to be needed again relatively soon.
The technology described herein makes a resource scaling decision based on average system load along a trend factor that indicates whether the load is increasing or decreasing. The use of the trend factor that predicts a direction of the system load makes the scalability decision as to whether to add or remove one or more resources more intelligent and more accurate, as the scalability determination is less likely to waste resources and/or prematurely remove distributed computing resources.
Note that as used herein, the terms “system” and “resource” with respect to load and scalability are generally interchangeable. For example, a system load can be measured by measuring the load (e.g., utilization) of a resource, and thus scalability can adjust the load by adding or removing resources. However, it is feasible to consider more than one resource as being scaled based on a measured resource load; e.g., if processor utilization is high and increasing, adding a processor can automatically add additional memory that the processor is likely to access. Indeed, some resources are typically a combination of resources; e.g., adding a physical server typically adds at least one processor and memory, and often adds disk storage. Moreover, as used herein, “number of computing resources” can be any one or more resources in any units; e.g., adding a disk drive resource can be to add a single physical disk drive regardless of its size, or can be to add a disk drive resource of X terabytes independent of how many actual physical disks are needed to provide that many terabytes, and so on.
As will be understood, the implementation(s) described herein are non-limiting examples, and variations to the technology can be implemented. For instances, many of the examples herein are based on the scaled distributed computing resources being a processor, e.g., the scaling determination is directed towards adding a processor, removing a processor or leaving the number of processors in the distributed computing resources as is until a next decision time. However, the technology described herein can provide benefits with any type of distributed computing resource that can have its load measured and be allocated/scaled on demand, including, but not limited to storage (any storage device/load in any units, e.g., memory in capacity or number of memory components, disk storage/load in number of disks or disk capacity), servers, physical or virtual machines, pods, containers, bandwidth (e.g., I/O and network), program instances, service instances, utilization of threads and so forth. As such, any of the embodiments, aspects, concepts, structures, functionalities, implementations and/or examples described herein are non-limiting, and the technology may be used in various ways that provide benefits and advantages in distributed computing and resource usage in general.
Reference throughout this specification to “one embodiment,” “an embodiment,” “one implementation,” “an implementation,” etc. means that a particular feature, structure, or characteristic described in connection with the embodiment/implementation is included in at least one embodiment/implementation. Thus, the appearances of such a phrase “in one embodiment,” “in an implementation,” etc. in various places throughout this specification are not necessarily all referring to the same embodiment/implementation. Furthermore, the particular features, structures, or characteristics may be combined in any suitable manner in one or more embodiments/implementations.
Aspects of the subject disclosure will now be described more fully hereinafter with reference to the accompanying drawings in which example components, graphs and operations are shown. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the various embodiments. However, the subject disclosure may be embodied in many different forms and should not be construed as limited to the examples set forth herein.
Turning to the drawings,
The distributed computing system 102 includes an operating system 106, such as a Linux®-based distrusted operating system. The operating system 106 manages the resources in use 108, and can add or remove resources to or from those resources in use 108 from a resource pool 110 or the like. Note that the resources in use 108 refer to those resources associated with a specific entity (e.g., program/system) that is using them, such as DELL/EMC's Pravega® open-source streaming storage system; if other entities are also using the distributed computing system 102, such other entities can have their own sets of allocated resources in use.
The automatic scaling component 104 comprises scaling logic 112, example operations of which are described herein with reference to 4-6. In general and as will be understood, the automatic scaling component 104 communicates with the operating system 106 of the distributed computing system 102 to scale up or scale down the resources in use 108 as deemed appropriate.
More particularly, in one or more implementations, the scaling logic 104, e.g., periodically or as otherwise needed, sends a load average report command (arrow 1) to the operating system 106. For each such command, the operating system 106 responds with a load average report (arrow 2). In this way, the scaling logic obtains the measured load average(s) over a prior period of time, which is one factor in determining whether to increase or decrease resources. Moreover, the reported load averages data 114 is maintained for use in trend prediction, by trend prediction logic 116, which as described herein is another factor (the trend factor) in determining whether to increase or decrease resources in use; arrow 3 represents scale-up/scale down commands for this purpose.
By way of example, in Linux®, “top” or “uptime” commands can be sent to obtain a load average report. As a more particular example, entering “uptime” on a Linux® system can return:
$ uptime
9:12:11 up 5 min, 3 users, load average: 0.56, 0.23, 0.68
In the above example, the three numbers in the report show the load average for the last minute (0.56), the last five minutes (0.23), and the last fifteen minutes (0.68).
Note that if only measuring CPU utilization, the load average definition sometimes can be ambiguous and simplistic in certain scenarios, as a more accurate system load average can calculate/combining utilization information for CPU, memory, bandwidth, disk load, and average utilization of threads, for example. Notwithstanding, for CPU-intensive systems, measuring only the CPU utilization is sufficient in many scenarios.
As another example, in the system, if the CPU processor number is 2, the “uptime” command returns
-
- $ uptime
- 9:30:10 up 23 min, 3 users, load average: 2.5, 2.3, 2.1
Prior solutions trigger workload scalability based on the load average threshold; for example, if the average load of the last five minutes is greater than a corresponding threshold value, the trigger event adds the resources, whereas if the average load of the last five minutes is less than a corresponding threshold value, the event is triggered to remove the resource. Note that the threshold level value for resource (CPU in this example) addition (L1) can be computed as the number of CPUs times a predetermined addition percentage such as eighty percent, and the threshold level value for resource removal (L0) can be computed as the number of CPUs times a predetermined removal percentage such as thirty percent.
An evaluation of the graph shows that at the point A0, the system load is increasing and thus removal of a resource as in the prior solutions is generally not desirable. Similarly, at the point A3, the system load is decreasing, whereby addition of a resource as in the prior solutions is generally not desirable.
The technology described herein uses this increasing or decreasing information, referred to herein as the trend or trend factor, in conjunction with the measured system (resource or resources) load to decide whether to add or remove a resource (which can be one or more resources). In this way, the following decisions are made with respect to the four points A0-A3 in
At the A0 point, A0<L0, but because the system load is increasing, the scaling component 104 does not remove the resource;
At the A1 point, A1<L0, and because the system load is decreasing, the scaling component 104 removes the resource;
At the A2 point, A2>L1, and D=1, because the system load is increasing, the scaling component 104 adds the resource; and
At the A3 point, A3>L1, but because the system load is decreasing, the scaling component 104 does not add the resource.
The trend factor can be represented by a value that indicates the increasing or decreasing direction (trend factor {right arrow over (D)}) of the system load, which can be represented as 1 (increasing) or −1 (decreasing), such that |{right arrow over (D)}|=1. If T represents a threshold, a system load threshold, the threshold with trend factor X can be abstracted into the formula:
X={right arrow over (D)}T,|{right arrow over (D)}|=1
As can be seen in
In this way, MACD can be used to predict load trends and directions, which thereby can generate resource addition and deletion signals. Note that the example shows a straightforward MACD trend determination, however other variations of MACD can be used, e.g., to wait a short time for a confirmed “crossover” to avoid being deceived and scale up or scale down resource usage too early, and so on. Further note that MACD is only one way to predict a trend, and other trend prediction technologies can be used instead of or in addition to MACD.
To summarize, based on the MACD trend prediction logic 116, the scaling logic obtains the trend factor {right arrow over (D)} value, and via the trend factor threshold, implements predictable automatic scaling to predict in advance whether to add or remove resources.
D=MACD ( )//return 1, or −1;
TA=CPU processor*80% for the resource addition threshold value;
TR=CPU processor*30% for the resource removal threshold value;
If (load>TA && D=1), then add the resource;
If (load<TR && D=−1), then remove the resource.
Operation 402 represents obtaining the system load, e.g., via a Linux® command, and operation 404 represents obtaining the trend factor, e.g., by calling the trend prediction logic 116 (a MACD function).
Operation 406 computes the threshold values, e.g., the number of CPUs in use times 0.8 for the addition threshold value, and the number of CPUs in use times 0.3 for the removal threshold value. Note that these multiplication factors can be varied based on empirical evidence, can differ based on type of resource and so on, however these factors have been found suitable for CPU scaling. In general, an addition threshold would use a factor greater than fifty percent, and a removal threshold computation would use a factor less than fifty percent.
Operation 408 evaluates whether the load value is greater than the resource addition threshold value. If not, the process continues to
If the load is below the resource addition threshold value, operation 502 of
If the load is below the resource removal threshold value, then operation 504 is performed to evaluate whether the trend factor indicates that the load is decreasing. If not, then again the current resource allocation is appropriate and nothing needs to be done at this time; thus the process ends until the next scalability decision is needed.
Operation 506 is an optional operation and represents, before removing a resource, whether the resource amount is at a minimum. For example, it may be beneficial to always have at least one CPU in use, even if idle, so that a user does not have to wait a frustrating amount of time to use a service. This is particularly true of resources that can take a long time to allocate; e.g., it can take on the order of minutes to spin up a virtual machine.
Thus, if below the resource removal threshold, and the trend factor indicates decreasing load, (and if evaluated, the resource is not at a minimum resource level), operation 508 removes the resource. The process resumes at the next resource scaling evaluation time, which can be triggered periodically, upon some event, at a certain time of day, and so forth.
Returning to operation 408 if the load value is greater than the resource addition threshold value, the process continues to
Similar to operation 506, operation 606 is an optional operation and represents, before adding a resource, whether the resource amount is at a maximum. For example, an enterprise may have reserved up to some amount of resources, and have to pay a large premium/penalty if going beyond that reserved number. Thus, before adding a resource beyond such a maximum, via optional operation 606 the resource addition can be bypassed, or possibly some other evaluation process can be invoked (e.g., whether this is a critical time-sensitive usage need, or can a user simply wait longer).
Thus, if above the resource addition threshold, and the trend factor indicates increasing load, (and if evaluated, the resource is not at a maximum resource level), operation 608 adds the resource. The process resumes at the next resource scaling evaluation time.
One or more aspects, generally exemplified in
Aspects can comprise, in response to the computing resource load value being less than a resource removal threshold value and the trend factor indicating that the computing resource load is decreasing, removing, by the system, a computing resource from the distributed computing system.
Aspects can comprise, in response to the computing resource load value being less than a resource removal threshold value, the trend factor indicating that the computing resource load is decreasing, and the computing resource not being at a minimum level, removing, by the system, a computing resource from the distributed computing system.
Aspects can comprise obtaining, by the system, the computing resource load value; the computing resource load value can comprise a reported resource load averaged over a prior interval time. Aspects can comprise obtaining, by the system, the trend factor based on a moving average convergence or divergence indicator of the computing resource load.
Aspects can comprise obtaining, by the system, the trend factor, wherein the trend factor indicates that the computing resource load is increasing when a moving average convergence divergence histogram is above a moving average convergence divergence baseline, and the trend factor indicates that the computing resource load is decreasing when the moving average convergence divergence histogram is below the moving average convergence divergence baseline.
Aspects can comprise obtaining, by the system, the computing resource addition threshold value based on multiplying a number of one or more computing resources by a predetermined addition percentage value.
The computing resource load can correspond to a number of one or more processors, and aspects can comprise obtaining, by the system, the computing resource addition threshold value based on multiplying the number of the one or more computing resources by about eighty percent.
Aspects can comprise, in response to the computing resource load value being less than a resource removal threshold value and the trend factor indicating that the computing resource load is decreasing, removing, by the system, a computing resource from the distributed computing system; the computing resource removal threshold value can be obtained by multiplying a number of computing resources by a predetermined removal percentage value. The computing resource load can correspond to a number of processors, and the computing resource removal threshold value can be obtained by multiplying the number of processors by about thirty percent.
The load value representative of the computing resource load can comprise a load average value determined over a prior time period. The load value representative of the computing resource load can correspond to measured processor utilization.
The trend factor can be based on a moving average convergence divergence indicator of the computing resource load. The addition threshold value can comprise a percentage value of greater than fifty percent multiplied by a number of one or more of the current computing resources currently in use. The removal threshold value can comprise a percentage value of less than fifty percent multiplied by a number of one or more of the current computing resources currently in use.
One or more aspects, such as implemented in a machine-readable storage medium, comprising executable instructions that, when executed by a processor, facilitate performance of operations, can be directed towards operations exemplified in
Further operations can comprise, obtaining the addition threshold value by multiplying the number of computing resources in use by an addition fractional value, and obtaining the removal threshold value by multiplying the number of computing resources in use by a removal fractional value. Obtaining the trend factor can comprise forecasting a system load trend using a result of a moving average convergence divergence prediction. Obtaining the load value can comprise requesting a report of a load average measured over a prior time period measured for the distributed computing system.
As can be seen, the technology described herein uses a load trend factor as part of determining whether to add or remove resource(s) to a distributed computing system. The technology thus is advantageous, including that it avoids unnecessary addition of unneeded resources or premature removal of resources likely to be needed soon.
The system 1000 also comprises one or more local component(s) 1020. The local component(s) 1020 can be hardware and/or software (e.g., threads, processes, computing devices). In some embodiments, local component(s) 1020 can comprise an automatic scaling component and/or programs that communicate/use the remote resources 1010 and 1020, etc., connected to a remotely located distributed computing system via communication framework 1040.
One possible communication between a remote component(s) 1010 and a local component(s) 1020 can be in the form of a data packet adapted to be transmitted between two or more computer processes. Another possible communication between a remote component(s) 1010 and a local component(s) 1020 can be in the form of circuit-switched data adapted to be transmitted between two or more computer processes in radio time slots. The system 1000 comprises a communication framework 1040 that can be employed to facilitate communications between the remote component(s) 1010 and the local component(s) 1020, and can comprise an air interface, e.g., Uu interface of a UMTS network, via a long-term evolution (LTE) network, etc. Remote component(s) 1010 can be operably connected to one or more remote data store(s) 1050, such as a hard drive, solid state drive, SIM card, device memory, etc., that can be employed to store information on the remote component(s) 1010 side of communication framework 1040. Similarly, local component(s) 1020 can be operably connected to one or more local data store(s) 1030, that can be employed to store information on the local component(s) 1020 side of communication framework 1040.
In order to provide a context for the various aspects of the disclosed subject matter,
In the subject specification, terms such as “store,” “storage,” “data store,” data storage,” “database,” and substantially any other information storage component relevant to operation and functionality of a component, refer to “memory components,” or entities embodied in a “memory” or components comprising the memory. It is noted that the memory components described herein can be either volatile memory or nonvolatile memory, or can comprise both volatile and nonvolatile memory, by way of illustration, and not limitation, volatile memory 1120 (see below), non-volatile memory 1122 (see below), disk storage 1124 (see below), and memory storage 1146 (see below). Further, nonvolatile memory can be included in read only memory, programmable read only memory, electrically programmable read only memory, electrically erasable read only memory, or flash memory. Volatile memory can comprise random access memory, which acts as external cache memory. By way of illustration and not limitation, random access memory is available in many forms such as synchronous random access memory, dynamic random access memory, synchronous dynamic random access memory, double data rate synchronous dynamic random access memory, enhanced synchronous dynamic random access memory, SynchLink dynamic random access memory, and direct Rambus random access memory. Additionally, the disclosed memory components of systems or methods herein are intended to comprise, without being limited to comprising, these and any other suitable types of memory.
Moreover, it is noted that the disclosed subject matter can be practiced with other computer system configurations, comprising single-processor or multiprocessor computer systems, mini-computing devices, mainframe computers, as well as personal computers, hand-held computing devices (e.g., personal digital assistant, phone, watch, tablet computers, netbook computers, . . . ), microprocessor-based or programmable consumer or industrial electronics, and the like. The illustrated aspects can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network; however, some if not all aspects of the subject disclosure can be practiced on stand-alone computers. In a distributed computing environment, program modules can be located in both local and remote memory storage devices.
System bus 1118 can be any of several types of bus structure(s) comprising a memory bus or a memory controller, a peripheral bus or an external bus, and/or a local bus using any variety of available bus architectures comprising, but not limited to, industrial standard architecture, micro-channel architecture, extended industrial standard architecture, intelligent drive electronics, video electronics standards association local bus, peripheral component interconnect, card bus, universal serial bus, advanced graphics port, personal computer memory card international association bus, Firewire (Institute of Electrical and Electronics Engineers 1394), and small computer systems interface.
System memory 1116 can comprise volatile memory 1120 and nonvolatile memory 1122. A basic input/output system, containing routines to transfer information between elements within computer 1112, such as during start-up, can be stored in nonvolatile memory 1122. By way of illustration, and not limitation, nonvolatile memory 1122 can comprise read only memory, programmable read only memory, electrically programmable read only memory, electrically erasable read only memory, or flash memory. Volatile memory 1120 comprises read only memory, which acts as external cache memory. By way of illustration and not limitation, read only memory is available in many forms such as synchronous random access memory, dynamic read only memory, synchronous dynamic read only memory, double data rate synchronous dynamic read only memory, enhanced synchronous dynamic read only memory, SynchLink dynamic read only memory, Rambus direct read only memory, direct Rambus dynamic read only memory, and Rambus dynamic read only memory.
Computer 1112 can also comprise removable/non-removable, volatile/non-volatile computer storage media.
Computing devices typically comprise a variety of media, which can comprise computer-readable storage media or communications media, which two terms are used herein differently from one another as follows.
Computer-readable storage media can be any available storage media that can be accessed by the computer and comprises both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer-readable storage media can be implemented in connection with any method or technology for storage of information such as computer-readable instructions, program modules, structured data, or unstructured data. Computer-readable storage media can comprise, but are not limited to, read only memory, programmable read only memory, electrically programmable read only memory, electrically erasable read only memory, flash memory or other memory technology, compact disk read only memory, digital versatile disk or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or other tangible media which can be used to store desired information. In this regard, the term “tangible” herein as may be applied to storage, memory or computer-readable media, is to be understood to exclude only propagating intangible signals per se as a modifier and does not relinquish coverage of all standard storage, memory or computer-readable media that are not only propagating intangible signals per se. In an aspect, tangible media can comprise non-transitory media wherein the term “non-transitory” herein as may be applied to storage, memory or computer-readable media, is to be understood to exclude only propagating transitory signals per se as a modifier and does not relinquish coverage of all standard storage, memory or computer-readable media that are not only propagating transitory signals per se. Computer-readable storage media can be accessed by one or more local or remote computing devices, e.g., via access requests, queries or other data retrieval protocols, for a variety of operations with respect to the information stored by the medium. As such, for example, a computer-readable medium can comprise executable instructions stored thereon that, in response to execution, can cause a system comprising a processor to perform operations, comprising determining a mapped cluster schema, altering the mapped cluster schema until a rule is satisfied, allocating storage space according to the mapped cluster schema, and enabling a data operation corresponding to the allocated storage space, as disclosed herein.
Communications media typically embody computer-readable instructions, data structures, program modules or other structured or unstructured data in a data signal such as a modulated data signal, e.g., a carrier wave or other transport mechanism, and comprises any information delivery or transport media. The term “modulated data signal” or signals refers to a signal that has one or more of its characteristics set or changed in such a manner as to encode information in one or more signals. By way of example, and not limitation, communication media comprise wired media, such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media.
It can be noted that
A user can enter commands or information into computer 1112 through input device(s) 1136. In some embodiments, a user interface can allow entry of user preference information, etc., and can be embodied in a touch sensitive display panel, a mouse/pointer input to a graphical user interface (GUI), a command line controlled interface, etc., allowing a user to interact with computer 1112. Input devices 1136 comprise, but are not limited to, a pointing device such as a mouse, trackball, stylus, touch pad, keyboard, microphone, joystick, game pad, satellite dish, scanner, TV tuner card, digital camera, digital video camera, web camera, cell phone, smartphone, tablet computer, etc. These and other input devices connect to processing unit 1114 through system bus 1118 by way of interface port(s) 1138. Interface port(s) 1138 comprise, for example, a serial port, a parallel port, a game port, a universal serial bus, an infrared port, a Bluetooth port, an IP port, or a logical port associated with a wireless service, etc. Output device(s) 1140 use some of the same type of ports as input device(s) 1136.
Thus, for example, a universal serial busport can be used to provide input to computer 1112 and to output information from computer 1112 to an output device 1140. Output adapter 1142 is provided to illustrate that there are some output devices 1140 like monitors, speakers, and printers, among other output devices 1140, which use special adapters. Output adapters 1142 comprise, by way of illustration and not limitation, video and sound cards that provide means of connection between output device 1140 and system bus 1118. It should be noted that other devices and/or systems of devices provide both input and output capabilities such as remote computer(s) 1144.
Computer 1112 can operate in a networked environment using logical connections to one or more remote computers, such as remote computer(s) 1144. Remote computer(s) 1144 can be a personal computer, a server, a router, a network PC, cloud storage, a cloud service, code executing in a cloud computing environment, a workstation, a microprocessor-based appliance, a peer device, or other common network node and the like, and typically comprises many or all of the elements described relative to computer 1112. A cloud computing environment, the cloud, or other similar terms can refer to computing that can share processing resources and data to one or more computer and/or other device(s) on an as needed basis to enable access to a shared pool of configurable computing resources that can be provisioned and released readily. Cloud computing and storage solutions can store and/or process data in third-party data centers which can leverage an economy of scale and can view accessing computing resources via a cloud service in a manner similar to a subscribing to an electric utility to access electrical energy, a telephone utility to access telephonic services, etc.
For purposes of brevity, only a memory storage device 1146 is illustrated with remote computer(s) 1144. Remote computer(s) 1144 is logically connected to computer 1112 through a network interface 1148 and then physically connected by way of communication connection 1150. Network interface 1148 encompasses wire and/or wireless communication networks such as local area networks and wide area networks. Local area network technologies comprise fiber distributed data interface, copper distributed data interface, Ethernet, Token Ring and the like. Wide area network technologies comprise, but are not limited to, point-to-point links, circuit-switching networks like integrated services digital networks and variations thereon, packet switching networks, and digital subscriber lines. As noted below, wireless technologies may be used in addition to or in place of the foregoing.
Communication connection(s) 1150 refer(s) to hardware/software employed to connect network interface 1148 to bus 1118. While communication connection 1150 is shown for illustrative clarity inside computer 1112, it can also be external to computer 1112. The hardware/software for connection to network interface 1148 can comprise, for example, internal and external technologies such as modems, comprising regular telephone grade modems, cable modems and digital subscriber line modems, integrated services digital network adapters, and Ethernet cards.
The above description of illustrated embodiments of the subject disclosure, comprising what is described in the Abstract, is not intended to be exhaustive or to limit the disclosed embodiments to the precise forms disclosed. While specific embodiments and examples are described herein for illustrative purposes, various modifications are possible that are considered within the scope of such embodiments and examples, as those skilled in the relevant art can recognize.
In this regard, while the disclosed subject matter has been described in connection with various embodiments and corresponding Figures, where applicable, it is to be understood that other similar embodiments can be used or modifications and additions can be made to the described embodiments for performing the same, similar, alternative, or substitute function of the disclosed subject matter without deviating therefrom. Therefore, the disclosed subject matter should not be limited to any single embodiment described herein, but rather should be construed in breadth and scope in accordance with the appended claims below.
As it employed in the subject specification, the term “processor” can refer to substantially any computing processing unit or device comprising, but not limited to comprising, single-core processors; single-processors with software multithread execution capability; multi-core processors; multi-core processors with software multithread execution capability; multi-core processors with hardware multithread technology; parallel platforms; and parallel platforms with distributed shared memory. Additionally, a processor can refer to an integrated circuit, an application specific integrated circuit, a digital signal processor, a field programmable gate array, a programmable logic controller, a complex programmable logic device, a discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. Processors can exploit nano-scale architectures such as, but not limited to, molecular and quantum-dot based transistors, switches and gates, in order to optimize space usage or enhance performance of user equipment. A processor may also be implemented as a combination of computing processing units.
As used in this application, the terms “component,” “system,” “platform,” “layer,” “selector,” “interface,” and the like are intended to refer to a computer-related entity or an entity related to an operational apparatus with one or more specific functionalities, wherein the entity can be either hardware, a combination of hardware and software, software, or software in execution. As an example, a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer. By way of illustration and not limitation, both an application running on a server and the server can be a component. One or more components may reside within a process and/or thread of execution and a component may be localized on one computer and/or distributed between two or more computers. In addition, these components can execute from various computer readable media having various data structures stored thereon. The components may communicate via local and/or remote processes such as in accordance with a signal having one or more data packets (e.g., data from one component interacting with another component in a local system, distributed system, and/or across a network such as the Internet with other systems via the signal). As another example, a component can be an apparatus with specific functionality provided by mechanical parts operated by electric or electronic circuitry, which is operated by a software or a firmware application executed by a processor, wherein the processor can be internal or external to the apparatus and executes at least a part of the software or firmware application. As yet another example, a component can be an apparatus that provides specific functionality through electronic components without mechanical parts, the electronic components can comprise a processor therein to execute software or firmware that confers at least in part the functionality of the electronic components.
In addition, the term “or” is intended to mean an inclusive “or” rather than an exclusive “or.” That is, unless specified otherwise, or clear from context, “X employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances.
While the invention is susceptible to various modifications and alternative constructions, certain illustrated implementations thereof are shown in the drawings and have been described above in detail. It should be understood, however, that there is no intention to limit the invention to the specific forms disclosed, but on the contrary, the intention is to cover all modifications, alternative constructions, and equivalents falling within the spirit and scope of the invention.
In addition to the various implementations described herein, it is to be understood that other similar implementations can be used or modifications and additions can be made to the described implementation(s) for performing the same or equivalent function of the corresponding implementation(s) without deviating therefrom. Still further, multiple processing chips or multiple devices can share the performance of one or more functions described herein, and similarly, storage can be effected across a plurality of devices. Accordingly, the invention is not to be limited to any single implementation, but rather is to be construed in breadth, spirit and scope in accordance with the appended claims.
Claims
1. A method, comprising:
- in a data storage system comprising data stored as objects and an object data structure, determining, by a system comprising a processor, whether an object is an embedded object based on a result of evaluating whether a total size of the object exceeds a first threshold size value; and in response to the object being determined to be an embedded object, storing object metadata of the object and object data of the object in the object data structure; and in response to the object being determined not to be the embedded object, storing at least part of the object in a chunk, and storing a pointer to the chunk in the object data structure.
2. The method of claim 1, wherein the object is determined to be the embedded object, and wherein the storing the object metadata of the object and the object data of the object in the object data structure comprises storing the object metadata of the object and the object data of the object in a single entry within the object data structure.
3. The method of claim 1, wherein the result is a first result, wherein the object is determined not to be the embedded object, and further comprising, determining whether the object is a normal object or an inflated object based on a second result of evaluating whether a size of the object metadata exceeds a second threshold size value.
4. The method of claim 3, wherein the object is a normal object, and wherein the storing at least the part of the object to the chunk comprises, storing the object metadata of the object in the object data structure in association with the pointer to the chunk, and storing the object data of the object in the chunk.
5. The method of claim 3, wherein the object is an inflated object, and wherein the storing at least the part of the object to the chunk comprises, storing at least part of the object metadata of the object in the chunk.
6. The method of claim 3, wherein the object is an inflated object, and wherein the storing at least the part of the object to the chunk comprises, storing the object metadata of the object in the chunk and storing the object data in the chunk.
7. A system, comprising:
- a data storage system configured to: store object-related information in an object table and in chunks, classify objects represented in the object table as embedded objects, normal objects or inflated objects via evaluations of size information of the objects, resulting in classified objects, in which corresponding object data size information for a classified object of the classified objects is based on object metadata size information of the classified object and object data size information of the classified object, and in response to the classified object being classified as an embedded object of the embedded objects, to store embedded object metadata of the embedded object and embedded object data of the embedded object in the object table; or in response to the classified object being classified as a normal object of the normal objects, to store normal object data of the normal object in a chunk, store a pointer to the chunk in the object table, and store at least some normal object metadata of the normal object in the object table in association with the pointer; or in response to the classified object being classified as an inflated object of the inflated objects, to store inflated object data of the inflated object in a chunk, store a pointer to the chunk in the object table, and store at least some inflated object metadata of the inflated object in the chunk.
8. The system of claim 7, wherein the corresponding object data size information of the classified object, comprising the object metadata size information in conjunction with the corresponding object data size information of the classified object, indicates a total object size below a threshold size value, and wherein the data storage system classifies the classified object as the embedded object.
9. The system of claim 7, wherein the corresponding object data size information of the classified object, comprising the object metadata size information, indicates an object metadata size below a threshold size value, and wherein the data storage system classifies the classified object as the normal object.
10. The system of claim 9, wherein the data storage system stores all or substantially all of the object metadata of the normal object in the object table in association with the pointer.
11. The system of claim 7, wherein the corresponding object data size information of the classified object, comprising the object metadata size information, indicates an object metadata size above a second threshold size value, and wherein the data storage system classifies the classified object as the inflated object.
12. The system of claim 11, wherein the data storage system stores all or substantially all of the object metadata of the inflated object in the chunk.
13. The system of claim 11, wherein the data storage system stores at least part of the object metadata of the inflated object in the object data structure.
14. The system of claim 7, wherein the corresponding object data size information of the classified object, comprising the object metadata size information in conjunction with the corresponding object data size information of the classified object, indicates a total object size above a first threshold size value, and further indicates an object metadata size below a second threshold size value, and wherein the data storage system classifies the classified object as the normal object.
15. The system of claim 14, wherein the first threshold size value is equal to the second threshold size value.
16. The system of claim 7, wherein the corresponding object data size information of the classified object, comprising the object metadata size information in conjunction with the corresponding object data size information of the classified object, indicates a total object size above a first threshold size value, wherein the corresponding object data size information of the classified object, comprising the object metadata size information, indicates an object metadata size above a second threshold size value, and wherein the data storage system classifies the classified object as the inflated object.
17. The system of claim 16, wherein the first threshold size value is equal to the second threshold size value.
18. A machine-readable storage medium, comprising executable instructions that, when executed by a processor, facilitate performance of operations, the operations comprising:
- obtaining an object for storage in a data storage system comprising data storage chunks, the data storage system comprising an object data structure comprising references to the data storage chunks; and
- evaluating object size information for the object, the object size information comprising object metadata size information of the object and object data size information of the object, for use in storing the object based on the evaluating, the evaluating and storing comprising: in response to the evaluating the object size information being determined to indicate a total object size below a first threshold value, storing object metadata of the object and object data of the object in the object data structure, in response to the evaluating the object size information being determined to indicate a total object size above the first threshold value, and the object metadata size information being determined to indicate that object metadata size is below a second threshold value, storing the object data of the object in a first chunk, storing a reference to the first chunk in the object data structure, and storing at least some of the object metadata of the object in the object data structure in association with the reference; or in response to the evaluating the object size information being determined to indicate a total object size above the first threshold value, and the object metadata size information being determined to indicate that object metadata size is above the second threshold value, storing the object data of the object in a second chunk, storing a reference to the second chunk in the object data structure, and storing at least some of the object metadata of the object in the second chunk.
19. The machine-readable storage medium of claim 18, wherein the evaluating the object size information indicates the total object size is above the first threshold value, and the object metadata size information indicates that object metadata size is below the second threshold value, and wherein the storing at least some of the object metadata of the object in the object data structure in association with the reference comprises storing all or substantially all of the object metadata of the object in the object data structure.
20. The machine-readable storage medium of claim 18, wherein the evaluating the object size information indicates the total object size is above the first threshold value, and the object metadata size information indicates that the object metadata size is above the second threshold value, and wherein the storing at least some of the object metadata of the object in the chunk comprises storing all or substantially all of the object metadata of the object in the chunk.