Detection of anomalous administrative actions

A method for monitoring includes defining a plurality of different types of administrative activities in a computer system. Each administrative activity in the plurality includes an action performed by one of the computers in the system that can be invoked only by a user having an elevated level of privileges in the system. The administrative activities performed by at least a group of the computers in the system are tracked automatically. Upon detecting that a given computer in the system has performed an anomalous combination of at least two of the different types of administrative activities, an action is initiated to inhibit malicious exploitation of the given computer.

Skip to: Description  ·  Claims  ·  References Cited  · Patent History  ·  Patent History
Description
FIELD OF THE INVENTION

The present invention relates generally to computer systems and networks, and particularly to apparatus, method and software for detecting malicious activity in computer networks.

BACKGROUND

In many computer and network systems, multiple layers of security apparatus and software are deployed in order to detect and repel the ever-growing range of security threats. At the most basic level, computers use anti-virus software to prevent malicious software from running on the computer. At the network level, intrusion detection and prevention systems analyze and control network traffic to prevent malware from spreading through the network.

In this latter category, for example, PCT International Publication WO 2013/014672, whose disclosure is incorporated herein by reference, describes a method and system for detecting anomalous action within a computer network. The method starts with collecting raw data from at least one probe sensor that is associated with at least one router, switch or server in the computer network. The raw data is parsed and analyzed to create meta-data from the raw data, and computer network actions are identified based on knowledge of network protocols. The meta-data is associated with entities by analyzing and correlating between the identified network actions. A statistical model of the computer network is created, for detection of anomalous network actions associated with the entities.

As another example, PCT International Publication WO 2014/111863, whose disclosure is incorporated herein by reference, describes a method for computer system forensics that includes receiving an identification of at least one host computer that has exhibited an anomalous behavior, in a computer network comprising multiple host computers. Respective images of the host computers in the network are assembled using image information collected with regard to the host computers. A comparison is made between at least one positive image of the at least one host computer, assembled using the image information collected following occurrence of the anomalous behavior, and one or more negative images assembled using the image information collected with respect to one or more of the host computers not exhibiting the anomalous behavior. Based on the comparison, a forensic indicator of the anomalous behavior is extracted from the positive and negative images.

SUMMARY

Embodiments of the present invention that are described hereinbelow provide methods, apparatus and software for detecting and inhibiting malicious activity in a computer network.

There is therefore provided, in accordance with an embodiment of the invention, a method for monitoring, which includes defining a plurality of different types of administrative activities in a computer system. Each administrative activity in the plurality includes an action performed by one of the computers in the system that can be invoked only by a user having an elevated level of privileges in the system. The administrative activities performed by at least a group of the computers in the system are tracked automatically. Upon detecting that a given computer in the system has performed an anomalous combination of at least two of the different types of administrative activities, an action is initiated to inhibit malicious exploitation of the given computer.

In a disclosed embodiment, the different types of the administrative activities are selected from a set of activities consisting of use of a secure shell protocol; login as an administrator; remote code execution; access to administrative resource shares; remote desktop operation on servers in the system; access to servers of servers in the system; and access to specified Web addresses that are reserved for administration. Additionally or alternatively, defining the plurality of different types of administrative activities includes specifying a set of administrative protocols, and tracking the administrative activities includes detecting uses of the administrative protocols in the set. Further additionally or alternatively the different types of the administrative activities include accessing non-existent network addresses and non-existent subnets.

In some embodiments, tracking the administrative activities includes establishing, for each of the computers in the group, a respective baseline level of the administrative activities, and detecting that a computer has performed an anomalous combination of the administrative activities includes detecting a deviation from the respective baseline level of the computer. Typically, establishing the respective baseline level includes identifying a set of the administrative activities performed by the computer over a first period of time, and detecting the deviation includes sensing, during a second period of time following the first period, that the computer has performed a type of administrative activity not in the identified set.

Additionally or alternatively tracking the administrative activities includes establishing a profile of the administrative activities performed over the group of the computers, and detecting that a computer has performed an anomalous combination of the administrative activities includes applying the profile in analyzing the administrative activities performed by the computer.

In a disclosed embodiment, establishing the profile includes assigning respective weights to the administrative activities responsively to respective frequencies of performance of the administrative activities by the computers in the system, such that the respective weights decrease as the respective frequencies increase, and analyzing the administrative activities includes computing a score by applying the weights to the administrative activities performed by the computer, and deciding that the combination of the administrative activities performed by the computer is anomalous if the score exceeds a predefined threshold. Typically, assigning the respective weights includes calculating a respective weight for each type of administrative activity in inverse proportion to a number of the computers performing the administrative activity.

There is also provided, in accordance with an embodiment of the invention, monitoring apparatus, which includes a memory, configured to store a definition of a plurality of different types of administrative activities in a computer system. A processor is configured to automatically track the administrative activities performed by at least a group of the computers in the system, and upon detecting that a computer in the system has performed an anomalous combination of at least two of the different types of administrative activities, to initiate an action to inhibit malicious exploitation of the given computer.

There is additionally provided, in accordance with an embodiment of the invention, a computer software product, including a non-transitory computer-readable medium in which program instructions are stored, which instructions, when read by a computer, cause the computer to store a definition of a plurality of different types of administrative activities in a computer system. The instructions cause the computer to automatically track the administrative activities performed by at least a group of the computers in the system, and upon detecting that a given computer in the system has performed an anomalous combination of at least two of the different types of administrative activities, to initiate an action to inhibit malicious exploitation of the given computer.

The present invention will be more fully understood from the following detailed description of the embodiments thereof, taken together with the drawings in which:

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 is a block diagram that schematically shows a computer system, in accordance with an embodiment of the present invention;

FIG. 2 is a flow chart that schematically illustrates a method for identifying anomalous administrative activities, in accordance with an embodiment of the invention;

FIG. 3 is a block diagram that schematically illustrates a method for profiling administrative activity in a network, in accordance with an embodiment of the invention; and

FIG. 4 is a plot that schematically shows levels of administrative activities over time, in accordance with an embodiment of the invention.

DETAILED DESCRIPTION OF EMBODIMENTS Overview

In almost every computer system (including computer networks), there are at least two different classes of users:

    • Standard users, who are allowed to operate the computer system and save documents, and may be able to install application programs and change personal settings, but cannot access settings and files belonging to the computer operating system.
    • Administrators, who have access to read and write any data in the system, add or remove any programs, and change operating system settings.
      In larger systems, there are frequently subdivisions within these classes. There remains, however, a clear division between “standard users” and “administrators” in nearly all multi-computer systems.

The distinction within the computer system between standard users and administrators (as well as between subdivisions of these classes) is expressed in terms of their respective levels of privilege. In the context of the present description and in the claims, “privilege” is defined as permission to perform a certain type of action in a computer system, subject to verification of the identity of the entity performing the action. In this sense, administrators have a higher level of privilege than standard users, since administrators can perform types of action that are barred to other users. Thus, in this context, administrative activities are defined as actions requiring elevated levels of privilege, which are normally permitted only to users who can present credentials verifying their identity as administrators.

Although under normal circumstances, elevated levels of privilege are permitted (as stated above) only to credentialed administrators, in many, if not most, cyber-attacks, the attacker attempts to gain an elevated level of privilege to which he or she is not entitled. This elevation of privilege may be gained, for example, by stealing or otherwise misappropriating the credentials of a legitimate administrator, or by exploiting a vulnerability in the system to take over the privileges of an existing software process. The elevated privileges enable the attacker to perform administrative actions by which he can steal, alter, or destroy certain data and/or operating components of the system under attack.

Although known types of attacks may be identified by their characteristic signatures, such techniques are ineffective in detecting new and targeted types of attacks, not to mention attacks based on misappropriation or misuse of legitimate administrative privileges. There is thus a need to recognize suspicious administrative activities, among the large body of administrative activities that occur every day, so that illegitimate activities can be promptly detected and inhibited.

Embodiments of the present invention that are described herein address this need by detecting anomalous administrative activities, which deviate from the normal pattern of administrative activities in a computer system. As explained above, administrative activities are defined in this context as actions performed by a computer in the computer system that can be invoked only by a user having an elevated level of privileges in the system. For the purposes of detecting possible security breaches, multiple different types of administrative activities are defined and tracked over all of the computers, or at least over a group of the computers, in the system being monitored. When a computer in the system performs an anomalous combination of two or more different types of these administrative activities, the combination is identified as suspicious, and action is initiated to inhibit malicious exploitation of these activities.

A combination of administrative activities is considered anomalous if it deviates from normal patterns of administrative activities in the system. It is difficult or impossible to define a “normal pattern” a priori, however, since such patterns can differ substantially from organization to organization and from system to system, and even among different computers in the same system. For example, not only will normal patterns of administrative activity differ between computers operated by standard users and those used by administrators, but even among different administrators the patterns of activity will vary depending on their respective roles and responsibilities.

Therefore, in some embodiments that are described below, a respective baseline level of administrative activities is established for each of the monitored computers. Anomalous combinations of administrative activities on any given computer are identified as deviations from the respective baseline level of that particular computer. The baseline level is established by learning the set of administrative activities performed by the given computer over a certain period of time, and a deviation is typically detected when the computer performs a type of administrative activity that is not in this set. In this manner, suspicious uses of legitimate administrative credentials may be detected, such as when the credentials of a given administrator, who normally performs a certain set of administrative activities, are suddenly used for a different type of administrative activity. This feature is important, since different administrators within a given system typically have different spheres of activity.

Furthermore, patterns of administrative activities typically vary substantially among different computer systems. To account for this variation, in some embodiments of the present invention, a profile is assembled of the administrative activities performed over the group of monitored computers in any given system, and this profile is applied in analyzing the administrative activities performed by the computers in the group. One way in which the profile can be used is in assigning respective weights to the administrative activities in a manner that reflects the respective frequencies of performance of these administrative activities by the computers in the system. The weights are then applied in computing a score for each computer, such that the combination of the administrative activities performed by that computer is considered anomalous if the weighted score exceeds a certain threshold. Specifically, administrative activities that are common in a given computer system receive low weights in that system, while unusual activities receive higher weights. In one embodiment, the respective weight for each type of administrative activity is in inverse proportion to the number of the computers performing the administrative activity during a certain period of time.

The combined use of computer-specific baselines and system-wide profiles enables embodiments of the present invention to detect suspicious administrative activities without prior definition of which activities should be considered anomalous. Rather, the anomaly detector itself learns the individual computer baselines and system-wide profile, and determines the degree to which any given combination of administrative activities should be considered anomalous. The present embodiments are therefore tolerant of changes in system configuration and even to errors of configuration, meaning that suspicious anomalies will be detected reliably, with a low rate of false alarms.

System Description

FIG. 1 is a block diagram that schematically shows a computer system 20, in which administrative activities are monitored in accordance with an embodiment of the present invention. System 20 comprises multiple personal computers and servers 24 (collectively referred to as host computers, or simply “hosts”), which are connected to a network 26, such as a local area network. Typically, each server 24 maintains a log 28 of actions performed by and on the server, such as logins, commands received and executed, and access to various resources on or via the server. Personal computers 22 may maintain similar logs (which are omitted from the figure for the sake of simplicity). Typically, system 20 is connected to public networks 30, such as the Internet, via a gateway 32, as is known in the art.

System 20 includes an authentication server 34, which authenticates and authorizes the users and computers in the system to perform requested activities, including specifically administrative activities. Authentication server 34 typically runs a directory service, which lists the authorized users and their respective privileges and credentials in system 20. Common directory services of this sort include, for example, the Microsoft® Active Directory service and Kerberos. Before performing privileged activities, hosts 22 and 24 must typically communicate with server 34 in order to receive the necessary permission. Server 34 authenticates and authorizes all users and computers in system 20 while enforcing security policies. Thus, when a user logs into a computer in system 20, server 34 checks the submitted password and determines whether the user is an administrator or a standard user. Server 34 typically maintains a log 36 of such activities and communications.

An anomaly detector 40 receives and processes information via network 26 from and about hosts 22, 24, and processes the information in order to detect anomalous administrative activities. Typically, anomaly detector 40 comprises a general-purpose computer, which includes a processor 42 and a memory 44 and is programmed in software to carry out the functions that are described herein. This software may be downloaded to the computer in electronic form, over a network, for example. Additionally or alternatively, the software may be stored on tangible, non-transitory computer-readable media, such as optical, magnetic, or electronic memory media. Further additionally or alternatively, at least some of the functions of processor 40 may be carried out by hard-wired or programmable digital logic circuits.

Although anomaly detector 40 is shown and described here for the sake of simplicity as a single, standalone unit, in practice the functions of the anomaly detector may be integrated with other software-based functions in a single server and/or distributed over a group of servers. In particular, the anomaly detector may be integrated in a server that also monitors and investigates other types of security threats in system 20. Such a server may also, for example, perform the anomaly detection functions described in the above-mentioned WO 2013/014672 and/or the forensic functions described in the above-mentioned WO 2014/111863.

Methods for Detection of Anomalous Activities

FIG. 2 is a flow chart that schematically illustrates a method for identifying anomalous administrative activities, in accordance with an embodiment of the invention. This method is described, for the sake of clarity and concreteness, with reference to anomaly detector 40 and the other elements of system 20. Alternatively, however, the techniques that are described herein may be implemented in substantially any computer with suitable processing capabilities and access to the necessary records of host activities. The steps in the method of FIG. 2 are shown in serial order for logical clarity, but in practice, the actions involved in the method are typically performed continually or at least repeated periodically, and may go on in parallel.

Processor 42 begins by collecting data regarding activities of hosts in the system being monitored (such as computers 22 and servers 24 and 34 in system 20), at a data collection step 50. The required data is typically collected, for example, by periodically reading information from logs 28 and 36. Processor 42 stores the collected data in memory 44 for further processing.

Within the collected data, processor 42 identifies administrative activities, at an activity sorting step 52. Generally speaking, as defined above, administrative activities are those that require elevated privileges in system 20. Practically speaking, a list of administrative activities is defined in advance, possibly with subsequent refinements from time to time, and is stored in memory 44. Processor 42 identifies activities that meet the criteria of the list. The list of administrative activities typically includes, for example, some or all of the following types of activities:

    • Use of a secure shell (SSH) protocol;
    • Login as an administrator;
    • Remote code execution;
    • Access to administrative resource shares;
    • Remote desktop operation on servers 24;
    • Access to servers of servers in system 20; and
    • Access to specified Web addresses on network 26 that are reserved for administration.

Additionally or alternatively, anomaly detector 40 may receive and store a list of administrative protocols and/or software programs that are used in system 20, and may then identify administrative activities by detecting uses of the administrative protocols and/or programs in the set defined by the list. In other words, any use of certain protocols or programs that are reserved in system 20 for administrators will be identified as an administrative activity.

As another example, simply registering a given host in an organizational unit of the directory service that is classified as an administrative unit in system 20 may be classed as an administrative activity.

As a further example, certain types of activities that are not necessarily administrative in character may be labeled as administrative by anomaly detector 40. For example, in performing networking configuration tasks, administrators may sometimes attempt to communicate with non-existent network addresses and non-existent subnets. Therefore, the anomaly detector may classify attempts to access non-existent network addresses and subnets as administrative activities.

Various methods of data analysis may be applied in recognizing the types of administrative activities defined above. For example, logins to administrator accounts may be detected by matching a suitable regular expression to the records in authentication log 36, such as the following regular expression:
regex=‘.*(admin|informationtechnology|information technology|systemnt|systemnt|administrators|(^|/)it($|/)).*’

As another example, remote code execution attempts in server logs 28 may be recognized by identifying remote procedure call (RPC) requests to a universally unique identifier (UUID) that is known as a service used to run code on remote hosts, such as the following UUIDs:

‘367ABB81:9844:35F1:AD32:98F038001003’

‘378E52B0:C0A9:11CF:822D:00AA0051E40F’

‘86D35949:83C9:4044:B424:DB363231FDOC’

‘1FF70682:0A51:30E8:076D:740BE8CEE98B’

‘0A74EF1C:41A4:4E06:83AE:DC74FB1CDD53’

Alternatively or additionally, processor 42 may recognize Server Message Block (SMB) requests with certain pipes that are known as services for running code on remote hosts, such as the ‘svcctl’ and ‘atsvc’ pipes.

For more precise classification, processor 42 may attempt to identify, for each such RPC or SMB request, a corresponding administrator authentication session in log 36 within some time limit, for example within one minute of the RPC or SMB request. If such an authentication session is found, processor 42 records the RPC or SMB request as an administrative activity.

For each host (computer 22 or server 24) in system 20, anomaly detector 40 computes an administrative activity baseline, at a baseline definition step 54. The baseline Bh of activities a for a given host h is defined as:
Bh=d∈T{a∈Adahas}
wherein Ad is the set of all activities ah performed by host h on day d that are classified as administrative activities as. An administrative activity as in this context is typically defined as a combination of an action (such as use of a particular protocol) and a target (such as the host to which a command using the protocol is directed). In other words, the same type of action directed to two different targets will be included as two distinct activities in the baseline; but repetition of the same activity within the given collection period will have no effect on the baseline. The baseline is taken over a training period T, which may be set, for example, to be twenty-one days. Typically, the baseline is updated periodically, for example, daily. Alternatively, longer or shorter collection and training periods may be defined.

Anomaly detector 40 also computes a profile of administrative activities of all of the monitored hosts 22, in system 20, at a profiling step 56. The profile indicates, for the current training period, how many hosts have performed each type of administrative activity that is being monitored.

FIG. 3 is a block diagram that schematically shows details of profiling step 56, in accordance with an embodiment of the invention. Aspects of the elements of FIG. 3 also play a role in identifying administrative activities at step 52.

Profiling begins from raw data 70 collected at step and results in compilation of an administrative activities profile 72. As explained above, multiple types of activities and information go into this profile:

    • User profiling 74 identifies computers 22 in system 20 from which users log in or otherwise identify themselves as administrators 76.
    • Host profiling 78 generates a list 80 of servers 24 and server activities in system 20. As explained above, certain server activities, such as remote code execution, are classified as administrative activities. The host profiling generates in particular a list 82 of servers of servers, i.e., servers that are normally accessed only by other servers (such as a database server that provides data to a Web server, for example). Any user access to a server in list 82 is typically identified as administrative.
    • Administrative host seeding 84 identifies hosts whose functions are essentially administrative. Within this cohort, administrative programs 88 may be identified, as well; and a “ping-pong” process may be applied to identify ports 90 used by programs 88 and further refine the list of administrative programs. On the other hand, certain hosts, such as servers that perform anti-virus functions, perform so many administrative activities with such regularity that it may be preferable to exclude them from the profile (and possibly not monitor them at all in the present context).
    • Pre-labeled activities 92—as explained above, activities that are not necessarily administrative in character but are labeled as administrative in the present context—are also added to profile 72.

Returning now to FIG. 2, during each monitoring period, anomaly detector 40 detects changes in the administrative activities performed by each monitored host h relative to its particular baseline Bh, at a change detection step 58. The monitoring period may be of any desired length, such as an hour, a day, or several days. Assuming the period to be one day, the change in administrative activities for host h on day d relative to its baseline is defined as Dh,d=Ah,d/Bh.

Based on the changes found for each host at step 58 and the profile assembled at step 56, processor 42 computes a weighted anomaly score for each host, at a weight computation step 60. The weight computed for each administrative activity depends on the frequency of performance of the activity over all the monitored hosts, and decreases as the frequency increases. Specifically, in the present embodiment, processor 42 typically counts the number of hosts that have performed each given administrative activity as during the current training period, and calculates a weight ws that is inversely proportional to this number:
ws=1/|{h∃{aahas}}|

The anomaly score for each host is then given by the sum of the weights of the administrative activities that have been performed by that host and are not included in the baseline activity set of the host:

S h , d = a D h , d w s
The weights ws are not necessarily constant over time, but may rather change with changes in the profile of administrative activities in system 20. Each non-baseline activity as (meaning a particular administrative action directed to a particular target, as explained above) is typically counted only once in this context per detection period. In other words, if the detection period is one day, and a given host initiates three SSH sessions on a particular server during the day in question, the weight assigned to this action is added only once to the score for the given host. Alternatively, other score computation algorithms may be applied.

Processor 42 compares the host scores to a predefined threshold, at an anomaly detection step 62. The threshold is chosen to give the optimal tradeoff between detection sensitivity and false alarm rate, and may be adjusted adaptively over time. Hosts whose anomaly scores exceed the threshold are flagged as possible sources of suspicious behavior. Anomaly detector 40 automatically invokes preventive action with respect to these hosts, at an activity inhibition step 64. This action may involve, at the simplest level, issuing an alarm to a system administrator, for example. Alternatively or additionally, anomaly detector 40 may quarantine hosts that exhibit suspicious behavior and prevent them entirely from communicating with other computers in system 20, or may instruct authentication server 34 to downgrade the privileges of these hosts.

On the other hand, as noted earlier, some hosts (particularly certain types of servers) routinely perform large numbers of administrative actions and may, in consequence, receive scores at step 60 that regularly exceed the threshold. In order to avoid repeated false alarms with respect to these hosts, anomaly detector 40 may place them in a whitelist that is excluded from further anomaly monitoring. For example, a host whose average score over a certain period, such as a week, is greater than a predefined threshold may be moved to the whitelist and ignored thereafter.

FIG. 4 is a plot that schematically shows levels of administrative activities over time in system 20, in accordance with an embodiment of the invention. The numbers of administrative activities shown on the vertical axis are arbitrary and are presented solely for the sake of illustration. An upper curve 100 shows the overall number of administrative activities performed by a given computer 22 in system 20 day by day. On the basis of these numbers and other factors, a threshold level 102 is established. This level may change over time, either manually or under administrative control, as the number and types of administrative activities by computer 22 changes over time.

A lower curve 104 shows the day-by-day number of administrative activities carried out by computer 22 that are not a part of its baseline activity set. In other words, curve 104 refers to a subset of the total administrative activity reflected by curve 100 that falls outside the baseline set. On most days, computer 22 performs no more than a single anomalous administrative activity of this sort. A peak 106, however, illustrates an abrupt increase in the number of these anomalous administrative activities performed by the computer on a certain day. Because peak 106 (or the score calculated on the basis of the peak at step 60) is above threshold level 102, anomaly detector 40 will detect the peak as suspicious and will initiate appropriate preventive action.

It will be appreciated that the embodiments described above are cited by way of example, and that the present invention is not limited to what has been particularly shown and described hereinabove. Rather, the scope of the present invention includes both combinations and subcombinations of the various features described hereinabove, as well as variations and modifications thereof which would occur to persons skilled in the art upon reading the foregoing description and which are not disclosed in the prior art.

Claims

1. A method for monitoring, comprising:

collecting, by a processor, data regarding activities of a plurality of computers in a computer system;
identifying in the plurality of computers, computers whose users identify as administrators;
identifying programs run by the computers identified as administrative;
generating a list of administrative activities of a plurality of different types, involving access to other computers, responsive to the identified programs and ports used by the identified programs;
identifying in the collected data, activities included in the generated list of administrative activities involving access to other computers;
assigning respective weights to the administrative activities in the generated list responsively to respective frequencies of performance of the administrative activities in the computer system, such that the respective weights decrease as the respective frequencies increase;
determining for each of the identified activities, a target computer of the activity;
establishing, for each of a plurality of computers in the computer system, a respective baseline including the targets accessed by the computer using one or more of the types of administrative activities in the generated list, and the specific types of administrative activities used in accessing each of the targets by the computer;
determining based on the collected data, the administrative activities performed by at least a group of the computers in the system, over a predetermined period, and the corresponding targets of the administrative activities;
determining pairs of the administrative activities and corresponding targets performed over the predetermined period by each of the computers in the group, not included in the respective established baseline of the computer;
computing, by the processor, a score for each of the computers, as a sum of the weights of the determined pairs of the administrative activities and corresponding targets performed by the computer during the predetermined period, that are not in the respective baseline, and deciding that the combination of the administrative activities performed by the computer is anomalous if the score exceeds a predefined threshold; and
upon detecting that a given computer in the system has performed an anomalous combination of administrative activities, initiating an action to inhibit malicious exploitation of the given computer,
wherein the different types of the administrative activities comprise accessing non-existent network addresses and non-existent subnets, and
wherein assigning the respective weights comprises calculating a respective weight for each type of administrative activity in inverse proportion to a number of the computers performing the administrative activity.

2. The method according to claim 1, wherein the different types of the administrative activities in the defined list, include:

access to administrative resource shares.

3. The method according to claim 1, wherein the plurality of different types of administrative activities comprises a set of administrative protocols, and wherein determining the administrative activities comprises detecting uses of the administrative protocols in the set.

4. The method according to claim 1, wherein establishing the respective baseline comprises identifying a set of the administrative activities performed by the computer over a first period of time.

5. The method according to claim 1, wherein determining the administrative activities comprises establishing a profile of the administrative activities performed over the group of the computers, and wherein detecting that a computer has performed an anomalous combination of the administrative activities comprises applying the profile in analyzing the administrative activities performed by the computer.

6. The method according to claim 1, wherein the different types of the administrative activities in the defined list, include: remote desktop operation on servers in the system.

7. The method according to claim 1, wherein the different types of the administrative activities in the defined list, include: access to servers-of-servers in the system.

8. The method according to claim 1, wherein the different types of the administrative activities in the defined list, include: access to specified Web addresses that are reserved for administration.

9. Monitoring apparatus, comprising:

a memory, configured to store a definition list of a plurality of different types of administrative activities involving access to other computers, in a computer system; and
a processor, which is configured to collect data regarding activities of a plurality of computers in the computer system, to identify in the plurality of computers, computers whose users identify as administrators, to identify programs run by the computers identified as administrative, to add to the definition list administrative activities, involving access to other computers, responsive to the identified programs and ports used by the identified programs, to identify in the collected data, activities included in the definition list, to assign respective weights to the administrative activities responsively to respective frequencies of performance of the administrative activities in the computer system, such that the respective weights decrease as the respective frequencies increase, to determine for each of the identified activities, a target computer of the activity, to establish, for each of a plurality of computers in the system, a respective baseline including the targets accessed by the computer using one or more of the types of administrative activities in the definition list, and the specific types of administrative activities used in accessing each of the targets by the computer, to determine based on the collected data, the administrative activities performed by at least a group of the computers in the system, over a predetermined period, and the corresponding targets of the administrative activities, to determine pairs of the administrative activities and corresponding targets performed over the predetermined period by each of the computers in the group, not included in the respective established baseline of the computer, to compute a score for each of the computers, as a sum of the weights of the determined pairs of the administrative activities and corresponding targets performed by the computer during the predetermined period, that are not in the respective baseline, and deciding that the combination of the administrative activities performed by the computer is anomalous if the score exceeds a predefined threshold, and upon detecting that a computer in the system has performed an anomalous combination of administrative activities, to initiate an action to inhibit malicious exploitation of the given computer,
wherein the different types of the administrative activities in the definition list comprise accessing non-existent network addresses and non-existent subnets, and
wherein the processor is configured to calculate the respective weight for each type of administrative activity in inverse proportion to a number of the computers performing the administrative activity.

10. The apparatus according to claim 9, wherein the different types of the administrative activities include:

use of a secure shell protocol;
login as an administrator;
remote code execution;
access to administrative resource shares;
remote desktop operation on servers in the system;
access to servers of servers in the system; and
access to specified Web addresses that are reserved for administration.

11. The apparatus according to claim 9, wherein the definition list of the plurality of different types of administrative activities comprises a specified set of administrative protocols, and wherein the processor is configured to determine the administrative activities by detecting uses of the administrative protocols in the set.

12. The apparatus according to claim 9, wherein the respective baseline comprises a set of the administrative activities performed by the computer over a first period of time.

13. The apparatus according to claim 9, wherein the processor is configured to establish a profile of the administrative activities performed over the group of the computers, and to apply the profile in analyzing the administrative activities performed by the computers in the group.

14. A computer software product, comprising a non-transitory computer-readable medium in which program instructions are stored, which instructions, when read by a computer, cause the computer to store a definition list of a plurality of different types of administrative activities involving access to other computers, in a computer system, to collect data regarding activities of a plurality of computers in the computer system, to identify in the plurality of computers, computers whose users identify as administrators, to identify programs run by the computers identified as administrative, to add to the definition list administrative activities, involving access to other computers, responsive to the identified programs and ports used by the identified programs, to identify in the collected data, activities included in the definition list, to assign respective weights to the administrative activities responsively to respective frequencies of performance of the administrative activities in the computer system, such that the respective weights decrease as the respective frequencies increase, to determine for each of the identified activities, a target computer of the activity, to establish, for each of a plurality of computers in the system, a respective baseline including the targets accessed by the computer using one or more of the types of administrative activities in the definition list, and the specific types of administrative activities used in accessing each of the targets by the computer, to determine based on the collected data, the administrative activities performed by at least a group of the computers in the system, over a predetermined period, and the corresponding targets of the administrative activities, to determine pairs of the administrative activities and corresponding targets performed over the predetermined period by each of the computers in the group, not included in the respective established baseline of the computer, to compute a score for each of the computers, as a sum of the weights of the determined pairs of the administrative activities and corresponding targets performed by the computer during the predetermined period, that are not in the respective baseline, and deciding that the combination of the administrative activities performed by the computer is anomalous if the score exceeds a predefined threshold, and upon detecting that a given computer in the system has performed an anomalous combination of administrative activities, to initiate an action to inhibit malicious exploitation of the given computer,

wherein the different types of the administrative activities in the definition list comprise accessing non-existent network addresses and non-existent subnets, and
wherein the instructions cause the computer to calculate the weight for each type of administrative activity in inverse proportion to a number of the computers performing the administrative activity.

15. The product according to claim 14, wherein the instructions cause the computer to establish a profile of the administrative activities performed over the group of the computers, and to apply the profile in analyzing the administrative activities performed by the computers in the group.

Referenced Cited
U.S. Patent Documents
5991881 November 23, 1999 Conklin et al.
6347374 February 12, 2002 Drake et al.
7178164 February 13, 2007 Bonnes
7694150 April 6, 2010 Kirby
7703138 April 20, 2010 Desai et al.
7752665 July 6, 2010 Robertson et al.
8429180 April 23, 2013 Sobel et al.
8490190 July 16, 2013 Hernacki et al.
8555388 October 8, 2013 Wang et al.
8677487 March 18, 2014 Balupari et al.
8925095 December 30, 2014 Herz et al.
8966625 February 24, 2015 Zuk et al.
9038178 May 19, 2015 Lin
9147071 September 29, 2015 Sallam
9342691 May 17, 2016 Maestas
9378361 June 28, 2016 Yen et al.
9386028 July 5, 2016 Altman
20030110396 June 12, 2003 Lewis et al.
20040117658 June 17, 2004 Klaes
20040210769 October 21, 2004 Radatti et al.
20040250169 December 9, 2004 Takemori et al.
20040260733 December 23, 2004 Adelstein et al.
20050262560 November 24, 2005 Gassoway
20050268112 December 1, 2005 Wang et al.
20060018466 January 26, 2006 Adelstein et al.
20060075462 April 6, 2006 Golan
20060075492 April 6, 2006 Golan et al.
20060075500 April 6, 2006 Bertman et al.
20060136720 June 22, 2006 Armstrong et al.
20060149848 July 6, 2006 Shay
20060161984 July 20, 2006 Phillips et al.
20060191010 August 24, 2006 Benjamin
20060282893 December 14, 2006 Wu et al.
20070072661 March 29, 2007 Lototski
20070214088 September 13, 2007 Graham
20070218874 September 20, 2007 Sinha et al.
20070226796 September 27, 2007 Gilbert et al.
20070283166 December 6, 2007 Yami et al.
20080005782 January 3, 2008 Aziz
20080016339 January 17, 2008 Shukla
20080016570 January 17, 2008 Capalik
20080104046 May 1, 2008 Singla et al.
20080104703 May 1, 2008 Rihn et al.
20080134296 June 5, 2008 Amitai et al.
20080198005 August 21, 2008 Schulak et al.
20080271143 October 30, 2008 Stephens et al.
20090007100 January 1, 2009 Field et al.
20090007220 January 1, 2009 Ormazabal et al.
20090157574 June 18, 2009 Lee
20090320136 December 24, 2009 Lambert et al.
20100107257 April 29, 2010 Ollmann
20100162400 June 24, 2010 Feeney et al.
20100197318 August 5, 2010 Petersen et al.
20100212013 August 19, 2010 Kim et al.
20100217861 August 26, 2010 Wu
20100268818 October 21, 2010 Richmond et al.
20100278054 November 4, 2010 Dighe
20100299430 November 25, 2010 Powers et al.
20110153748 June 23, 2011 Lee et al.
20110185421 July 28, 2011 Wittenstein et al.
20110214187 September 1, 2011 Wittenstein et al.
20110225650 September 15, 2011 Margolies
20110247071 October 6, 2011 Hooks et al.
20110265011 October 27, 2011 Taylor et al.
20110270957 November 3, 2011 Phan et al.
20110302653 December 8, 2011 Frantz et al.
20110321175 December 29, 2011 Slater
20120079596 March 29, 2012 Thomas et al.
20120102359 April 26, 2012 Hooks
20120136802 May 31, 2012 Mcquade et al.
20120137342 May 31, 2012 Hartrell et al.
20120143650 June 7, 2012 Crowley et al.
20120191660 July 26, 2012 Hoog
20120222120 August 30, 2012 Rim et al.
20120233311 September 13, 2012 Parker et al.
20120275505 November 1, 2012 Tzannes et al.
20120331553 December 27, 2012 Aziz et al.
20130097706 April 18, 2013 Titonis et al.
20130111211 May 2, 2013 Winslow et al.
20130196549 August 1, 2013 Sorani
20130298243 November 7, 2013 Kumar et al.
20140165207 June 12, 2014 Engel et al.
20140196115 July 10, 2014 Pelykh
20140325643 October 30, 2014 Bart et al.
20150040219 February 5, 2015 Garraway
20150047032 February 12, 2015 Hannis et al.
20150121461 April 30, 2015 Dulkin et al.
20150195300 July 9, 2015 Adjaoute
20150200821 July 16, 2015 Sade
20150264069 September 17, 2015 Beauchesne et al.
20150286819 October 8, 2015 Coden
20150304346 October 22, 2015 Kim
20160191918 June 30, 2016 Lai et al.
20160306965 October 20, 2016 Iyer
20160315954 October 27, 2016 Peterson
Foreign Patent Documents
0952521 October 1999 EP
2056559 May 2009 EP
Other references
  • Shulman el al., “Top Ten Database Security Threats How to Mitigate the Most Significant Database Vulnerabilities” white paper, Imperva Inc (2006); pp. 1-14.
  • European Application # 14741051.8 Search Report dated Jul. 15, 2016.
  • Bilge et at., “Disclosure: Detecting Botnet Command and Control Servers Through Large-Scale NetFlow Analysis”, ACSAC, 10 Pages, Dec. 3-7, 2012.
  • Blum., “Combining Labeled and Unlabeled Data with Co-Training”, Carnegie Mellon University, Research Showcase @ CMU, Computer Science Department, 11 pages, Jul. 1998.
  • Felegyhazi et al., “On the Potential of Proactive Domain Blacklisting”, LEET'10 Proceedings of the 3rd USENIX Conference on Large-scale exploits and emergent threats, 8 pages, San Jose, USA, Apr. 27, 2010.
  • Frosch., “Mining DNS-related Data for Suspicious Features”, Ruhr Universitat Bochum, Master'sThesis, 88 pages, Dec. 23, 2011.
  • Bilge at al., “Exposure: Finding Malicious Domains Using Passive DNS Analysis”, NDSS Symposium, 17 pages, Feb. 6-9, 2011.
  • Gross et al., “FIRE: Finding Rogue nEtworks”, Annual Conference on Computer Security Applications (ACSAC'09), 10 pages, Dec. 7-11, 2009.
  • Markowitz, N., “Bullet Proof Hosting: A Theoretical Model”, Security Week, 5 pages, Jun. 29, 2010, downloaded from http://www.infosecisland.com/blogview/4487-Bullet-Proof-Hosting-A-Theoretical-Model.html.
  • Konte et al., “ASwatch: An AS Reputation System to Expose Bulletproof Hosting ASes”, SIGCOMM , pp. 625-638, Aug. 17-21, 2015.
  • Markowitz, N., “Patterns of Use and Abuse with IP Addresses”, Security Week, 4 pages, Jul. 10, 2010, downloaded from http://infosecisland.com/blogview/5068-Patterns-of-Use-and-Abuse-with-IP-Addresses.html.
  • Wei et al., “Identifying New Spam Domains by Hosting IPs: Improving Domain Blacklisting”, Department of Computer and Information Sciences, University of Alabama at Birmingham, USA, 8 pages, Dec. 8, 2010.
  • Goncharov,M., “Criminal Hideouts for Lease: Bulletproof Hosting Services”, Forward-Looking Threat Research (FTR) Team, A TrendLabsSM Research Paper, 28 pages, Jul. 3, 2015.
  • International Patent Application # PCT/IL2012/050272 Search Report dated Nov. 28, 2012.
  • Tier-3 Pty Ltd, “Huntsman Protector 360”, Brochure, 2 pages, Apr. 1, 2010.
  • Tier-3 Pty Ltd, “Huntsman 5.7 The Power of 2”, Brochure, 2 pages, Oct. 8, 2012.
  • Light Cyber Ltd, “LightCyber Magna”, 3 pages, year 2011.
  • International Application # PCT/IB2014/058299 Search report dated May 25, 2014.
  • European Application # 12817760.7 Search Report dated Apr. 1, 2015.
  • Mumcuoglu et al., U.S. Appl. No. 14/758,966, filed Jul. 2, 2015.
  • U.S. Appl. No. 14/234,165 Office Action dated Mar. 25, 2015.
  • U.S. Appl. No. 14/234,165 Office Action dated Sep. 22, 2015.
  • European Application # 16167582.2 Search Report dated Aug. 2, 2016.
  • U.S. Appl. No. 14/758,966 Office Action dated Oct. 26, 2016.
  • Niksun, “Network Intrusion Forensic System (NIFS) for Intrusion Detection and Advanced Post Incident Forensics”, Whitepaper, 12 pages, Feb. 15, 2010.
  • U.S. Appl. No. 15/286,674 Office Action dated Apr. 12, 2017.
  • U.S. Appl. No. 15/286,643 Office Action dated Mar. 17, 2017.
  • U.S. Appl. No. 15/075,343 Office Action dated Mar. 30, 2017.
  • U.S. Appl. No. 14/758,966 Office Action dated May 19, 2017.
  • EP Application # 16167582.2 office action dated Nov. 17, 2017.
Patent History
Patent number: 10075461
Type: Grant
Filed: May 31, 2015
Date of Patent: Sep 11, 2018
Patent Publication Number: 20170054744
Assignee: PALO ALTO NETWORKS (ISRAEL ANALYTICS) LTD. (Tel Aviv)
Inventors: Michael Mumcuoglu (Jerusalem), Giora Engel (Mevasseret-Zion), Yaron Neuman (Zoran-Kadima), Eyal Firstenberg (Ramat HaSharon)
Primary Examiner: Harunur Rashid
Application Number: 14/726,539
Classifications
Current U.S. Class: Access Control Or Authentication (726/2)
International Classification: H04L 29/06 (20060101);