[go: up one dir, main page]

US20130318616A1 - Predicting attacks based on probabilistic game-theory - Google Patents

Predicting attacks based on probabilistic game-theory Download PDF

Info

Publication number
US20130318616A1
US20130318616A1 US13/487,774 US201213487774A US2013318616A1 US 20130318616 A1 US20130318616 A1 US 20130318616A1 US 201213487774 A US201213487774 A US 201213487774A US 2013318616 A1 US2013318616 A1 US 2013318616A1
Authority
US
United States
Prior art keywords
network
attacker
nodes
event information
potential targets
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/487,774
Inventor
Mihai Christodorescu
Dmytro Korzhyk
Reiner Sailer
Douglas L Schales
Marc Ph Stoecklin
Ting Wang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US13/487,774 priority Critical patent/US20130318616A1/en
Publication of US20130318616A1 publication Critical patent/US20130318616A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/14Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
    • H04L63/1408Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic by monitoring network traffic
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/55Detecting local intrusion or implementing counter-measures
    • G06F21/552Detecting local intrusion or implementing counter-measures involving long-term monitoring or reporting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0637Strategic management or analysis, e.g. setting a goal or target of an organisation; Planning actions based on goals; Analysis or evaluation of effectiveness of goals
    • G06Q10/06375Prediction of business process outcome or impact based on a proposed change
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/20Network architectures or network communication protocols for network security for managing network security; network security policies in general

Definitions

  • the present invention relates to computer security and, more particularly, to modeling instances and targets for in-progress attacks using probabilistic game theory.
  • an entity may be a web server, an SQL server, a user terminal, etc.
  • an entity may be a web server, an SQL server, a user terminal, etc.
  • a strongly connected network the removal of a small number of connections will not partition the network into isolated parts.
  • detection and response systems do not provide adequate insight to system operators as to how best to respond to a strategic attacker.
  • targets are numerous and easily reachable, making existing approaches that assume a small target set impractical to use.
  • a method for determining cyber-attack targets includes collecting and storing network event information from a plurality of sensors to extract information regarding an attacker; forming an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets; calculating a likelihood for each of the plurality of paths using a processor; and calculating a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker.
  • a method for determining cyber-attack targets includes collecting and storing network event information from a plurality of sensors to extract information regarding an attacker; forming an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets; calculating a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker using a processor; determining a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and removing the determined network graph edge.
  • a method for determining cyber-attack targets includes collecting and storing network event information from a plurality of sensors to extract information regarding an attacker; forming an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets; calculating a likelihood for each of the plurality of paths using a processor; and calculating a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker; calculating a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker; determining a network graph edge to remove which minimizes a defender's expected uncertainty over the potential targets; and removing the determined network graph edge.
  • a system for determining cyber-attack target includes a network monitor module configured to collect network event information from sensors in one or more network nodes; a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a likelihood for each of the plurality of paths, and to calculate a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker.
  • a system for determining cyber-attack targets includes a network monitor module configured to collect network event information from sensors in one or more network nodes; a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker, and to determine a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and a network management module configured to remove the determined network graph edge.
  • a system for determining cyber-attack target includes a network monitor module configured to collect network event information from sensors in one or more network nodes; a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a likelihood for each of the plurality of paths, and to calculate a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker, to calculate a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker, and to determine a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and a network management module configured to remove the determined network graph edge.
  • FIG. 1 is a diagram of a network graph and an associated attack tree according to the present principles
  • FIG. 2 is an alternative representation of the mapping between the network graph and the associated attack tree according to the present principles
  • FIG. 3 is a block/flow diagram of a method for determining an attack target according to the present principles
  • FIG. 4 is a block/flow diagram of an alternative method for determining an attack target according to the present principles
  • FIG. 5 is a graph showing a reduction in entropy after removing a network graph edge in accordance with the present principles
  • FIG. 6 is a diagram of a system for determining an attack target according to the present principles.
  • the present principles employ game theory to predict attacker targets. Using a probabilistic model of attacker behavior, the interactions between a network defender and attacker are modeled, allowing the defender to anticipate future steps of the attack and identify the most likely attack targets based on the observed network events.
  • the present principles use attack scenario trees which represent the possible sequences of high-level attack steps that can be executed at different nodes of the network. This approach differs from the attack-response trees used previously, which represent attack steps within a single network host. Attack scenario trees can be constructed based on past incident reports.
  • the interaction between the defender and the attacker is modeled as a two-player Stackelberg game.
  • the defender can use the model to further decrease uncertainty about attack target predictions by blocking specific network paths (and indirectly any attack steps that traverse those paths) and influencing the attacker to reveal their intentions while conducting the attack. This allows defenders the benefit of proactively blocking future attack steps.
  • a given network 100 may be represented by a graph of interconnected nodes, with each node being a device or point of vulnerability on the network, and each edge representing a communication link between two nodes. Each node also has one or more associated vulnerabilities, which an attacker may exploit to compromise the security of the node. Combining the vulnerabilities with the edges of the graph provide structure for the attack scenario tree.
  • the nodes of the attack scenario tree represent high-level steps of a potential attack on a defended network.
  • an attacker 102 may approach network 100 from one of three externally available systems: web server 104 , file server 106 , and an email or web client 108 .
  • Each of these points of attack has an associated vulnerability.
  • web server 104 may be vulnerable to an SQL injection attack 116
  • file server 106 may be vulnerable to a respective file server attack 114 .
  • Compromising either of these nodes may give access to active directory (AD) server 110 , allowing the attacker 102 to gain access to AD credentials 120 .
  • the attacker 102 can then use the AD credentials to install a remote access tool 122 on a target device 112 .
  • AD active directory
  • any given node may be an AND node, where all of the nodes leading up to it must be reached before accessing the AND node, an OR node where any one input node may be reached, or the nodes may implement any other condition or combination of conditions.
  • the present principles allow a defender to monitor the attack in-progress and provide probabilistic information regarding likely attack paths and targets. By blocking the attacker's access to particular connections, the attack scenario tree can be trimmed and certainty regarding the attacker's goals can be increased.
  • Each node v ⁇ V belongs to a certain type ⁇ (v), where ⁇ :V ⁇ .
  • the node types incorporate the information about node vulnerabilities that can be exploited by the attacker 102 .
  • the attack scenario trees are thus constructed of elements of the type set ⁇ .
  • Cyber attacks can be assessed from either the point of view of the defender or from the point of view of the attacker.
  • the attacks are paths in the network graph G from the attacker's starting point v a ⁇ V to a target t ⁇ V.
  • an attack is a path from one of the leaves to the root of an attack scenario tree s ⁇ S. Since s is composed of node types ⁇ (v) and not of specific nodes, each path in an attack scenario tree s can correspond to multiple paths in G as long as the sequence of node types in G matches the types in scenario s.
  • I(s) is the set of instantiations of the attack scenario s.
  • Each node is identified by a name and a type.
  • an attacker starting at node 202 can proceed to one of three nodes 204 .
  • the attacker may then take different paths to reach one of the three target nodes 206 .
  • the attacker always has control over the starting node 202 v a ⁇ V.
  • the set of active nodes a ⁇ V which are controlled by the attacker expands until t ⁇ A.
  • This process of attacker's expansion over the nodes in V corresponds to a simultaneous expansion of the set ⁇ of active node types in the attack scenario tree until includes the root of the tree. Inferring a probability distribution over the possible sets A and helps in predicting the attack targets.
  • a Stackelberg two-player game models strategic interaction between two intelligent agents, designated the leader and the follower.
  • Each player has a finite set of actions to choose from.
  • the leader's set of actions is marked as A l
  • the follower's set of actions is marked as A f .
  • a pair of actions (a l ,a f ) chosen by the players is called the outcome of the game.
  • the players' utilities are functions of the outcome.
  • the leader's utility function is denoted by u f (a l ,a f ) and the follower's utility function by u l (a l ,a f ).
  • the game proceeds as follows.
  • the leader chooses (or commits to) a mixed strategy having a probability distribution over the actions in A l .
  • the follower observes the distribution and chooses a best-response which, generally, can be a probability distribution over A f .
  • a best-response which, generally, can be a probability distribution over A f .
  • one of the optimal best-responses is a degenerate distribution.
  • an optimal mixed strategy for the leader can be computed in polynomial time using linear programming techniques.
  • the follower's preferences can be represented by choosing an action for each distribution over the leader's actions, instead of using the follower's utility function u f .
  • the follower's preferences are written as a mapping f: ⁇ (A l ) ⁇ A f , where the operator ⁇ (.) denotes a set of distributions over a given finite set.
  • the present principles provide an extension to the described Stackelberg model in which the follower's preferences are described by a mapping f: ⁇ (A l ) ⁇ (A f ). In other words, if the leader commits to a mixed strategy s l ⁇ (A l ) then the follower plays a mixed strategy f(s l ).
  • This extension is called herein a probabilistic Stackelberg model.
  • a probabilistic Stackelberg model can be used to describe irrational behavior of the follower.
  • the present principles provide a probabilistic Stackelberg model in which the function f is a linear mapping from the vector of probabilities describing the leader's mixed strategy to the vector of probabilities describing the follower's mixed strategy.
  • Table 1 shows the potential increase in utility that the leader can achieve by considering a probabilistic follower model rather than assuming that the follower is perfectly rational and optimizing a known utility function is shown.
  • the leader is the row player and the follower is the column player.
  • the leader has two strategies U,D, and the follower has two strategies L,R.
  • Each cell in the table shows the leader's and the follower's utility for the corresponding choice of actions. If the follower is perfectly rational and always chooses the action that maximizes its utility, the optimal strategy to commit to for the leader is
  • the follower best-responds with L if the follower plays U with probability 0.5 or higher, and the follower plays D otherwise.
  • the leader can get an even higher utility by deviating to play D more frequently, because the follower will still play R with a relatively high probability after such deviation.
  • This example demonstrates the potential benefits of using a probabilistic follower model, whether derived from the follower's utility function or defined directly as a probability function on the set of the defender's actions.
  • Block 302 monitors such network events.
  • Block 306 estimates a probability distribution over the sets of active nodes A and active node types . Toward this end, block 306 computes heuristical likelihood scores l for each instantiation path ⁇ in I(s) as follows
  • l ⁇ ( ⁇ ) max q ⁇ 1 ⁇ ⁇ ... ⁇ ⁇ q m , r 1 ⁇ ⁇ ... ⁇ ⁇ r m ⁇ ⁇ i ⁇ ⁇ p ⁇ ( o q i ⁇ ⁇ r i ) ⁇ ⁇ j ⁇ r 1 , ⁇ ... ⁇ , r m ⁇ ⁇ p _ ⁇ ( ⁇ j )
  • ⁇ r i ) is the probability of making an observation o q i when the node ⁇ r i is attacked
  • p ( ⁇ j) is the probability of the node ⁇ j being attacked without triggering any observations.
  • block 308 computes a probability distribution over the targets t ⁇ V under the assumption that the attacker will follow one of the attack scenario trees in S according to the estimated distribution over the active node types and the active network nodes A.
  • the distribution over the attack targets is estimated using Monte Carlo simulation:
  • A] is estimated by simulating the attack steps according to the corresponding attack scenario tree starting with active node types .
  • the method of FIG. 3 is repeated each time an attacker advances the attack to a new node.
  • the attacker chooses a network node v adjacent to the set of active nodes A such that the type ⁇ (v) matches one of the possible next steps in the current state of the attack scenario tree.
  • Such attack actions are simulated until the set A expands to include one of the targets t, at which point t is assumed to be the intended target of the attack in the current run of the simulation.
  • the frequency with which target t is attacked in the simulations is the estimate of P[t
  • aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
  • a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider an Internet Service Provider
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • the defender can modify the network 100 in order to force the attacker 102 to reveal the intended target 112 of the attack.
  • one way for the defender to react to an ongoing cyberattack is to block network connections over a certain port.
  • the defender has the ability to block a network edge in response to the observed network events. Since there are likely multiple ways from the attacker's starting node 202 to the any target 206 in a realistic computer network 100 , blocking one edge will not prevent the attack. However, the defender may improve the future predictions about the attacker's intended target by carefully choosing a single graph edge to block.
  • block 404 Given the observed network events, collected at block 402 , block 404 computes a posterior probability distribution over the attacker's state (A, ).
  • the optimal edge e ⁇ E to block is the one that minimizes the defender's expected future uncertainty over the attack targets given the history of the observed network events.
  • Block 408 implements that action by, e.g., disabling the network connection represented by edge e.
  • the attacker has only reached node s 202 , and all attack paths going from the bottom of the graph to the top are possible. Additionally, the attacker chooses the next possible node to attack uniformly at random. At first, the probability distribution over the targets is (1/3,1/3,1/3). Note that the distribution will remain the same if the attacker chooses node b as the next node to attack. However, if the attacker chooses node a, then it is certain that t 1 is the target of the attack. If the attacker chooses node c, targets t 2 and t 3 are two equally likely targets.
  • FIG. 5 a graph showing expected entropy for two scenarios is shown.
  • the vertical axis shows the calculated entropy, while the horizontal axis shows which step of the attack is being considered.
  • the expected entropy is plotted with a solid line.
  • the entropy can be reduced, however, if the possibility of the attacker choosing node b is removed. Removing the edge between s and b results in a reduction of the expected entropy after the next attack step. This reduced entropy with a dashed line on the same plot.
  • step 2 entropy is zero for both plots because the attacker has reached the target t.
  • the present principles provides a significant improvement in terms of the defender's uncertainty as to the attacker's potential targets. Moreover, by blocking paths in concordance with the probabilistic Stackelberg model of the present principles, the defender is able to further reduce the uncertainty by influencing the attacker's next move. Furthermore, the present principles are scalable, with good performance in attack paths of length 3 , 4 , and 5 —it should be noted that practical attack paths rarely exceed four steps in length. Simulations run with over 1000 nodes and with attack path lengths of five were able to complete in a matter of seconds, making the present principles practical for real-world application.
  • a defender employs a defender administration terminal 602 to observe and respond to an attacker's actions.
  • the defender administration terminal 602 may be a stand-alone device, or may be part of a more general purpose administration terminal.
  • the defender terminal 602 communicates with sensors 604 distributed throughout a network 100 to collect information regarding potential intrusions.
  • the sensors may be incorporated in any node in a network 100 and may monitor any process or component to detect irregular activities. Multiple sensors 604 may be incorporated in a single node. Sensors 604 may provide detailed information regarding the attacker and its activities, may report on any irregular activity, or may simply represent logs for all activities in the system, said logs to be analyzed by the defender terminal 602 .
  • the sensors 604 may include software installed on nodes to monitor system calls and user activities, or alternatively be standalone devices such as traffic sniffers configured to detect particular suspicious forms of network traffic to or from monitored nodes.
  • the defender administration terminal 602 processes the data provided by sensors 604 and determines likely targets for the attacker.
  • the defender terminal 602 also calculates an optimal defender response to reduce the uncertainty in the attacker's targets.
  • Defender terminal 602 includes a processor 608 and memory 610 to collect and utilize the sensor data using network monitor 614 .
  • the network monitor 614 collects data from sensors 604 and processes that data from potentially heterogeneous sources into a usable form.
  • the network monitor may, for example, parse logs provided by sensors 604 to find suspicious or abnormal entries.
  • the processor 608 uses the sensor data provided by network monitor 614 and stored in memory 610 to produce the most probable targets and response.
  • the network control module 612 executes the response using a network management interface 606 .
  • the network management 606 interface may represent any appropriate form of network management, including for example a simple network management protocol (SNMP) device.
  • SNMP simple network management protocol
  • the defender administration terminal 602 can disconnect links in a network 100 , or take similar network-level measures that prevent the attacker to proceed along the chosen network links (e.g., the defender could choose to enable a firewall system on that link, instead of disconnecting the link).
  • the changed network topology forces the attacker along different paths, and the attacker's response to the defender's action substantially reduces the uncertainty regarding the attacker's intentions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Computer Security & Cryptography (AREA)
  • Human Resources & Organizations (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Economics (AREA)
  • Educational Administration (AREA)
  • Strategic Management (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computing Systems (AREA)
  • General Business, Economics & Management (AREA)
  • Game Theory and Decision Science (AREA)
  • Development Economics (AREA)
  • Marketing (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Tourism & Hospitality (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

Systems for determining cyber-attack target include a network monitor module configured to collect network event information from sensors in one or more network nodes; a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a likelihood for each of the paths, to calculate a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker, to calculate a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker, and to determine a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and a network management module configured to remove the determined network graph edge.

Description

    RELATED APPLICATION INFORMATION
  • This application is a Continuation application of pending U.S. patent application Ser. No. 13/478,290 filed on May 23, 2012, incorporated herein by reference in its entirety.
  • BACKGROUND
  • 1. Technical Field
  • The present invention relates to computer security and, more particularly, to modeling instances and targets for in-progress attacks using probabilistic game theory.
  • 2. Description of the Related Art
  • A large increase in the frequency of cybersecurity attacks has prompted industry and academia to find new ways to respond to the threat. Defensive mechanisms have been proposed in an attempt to detect and prevent attackers from reaching their targets, e.g., servers that store high-value data. In practice, large networks can have hundreds of high-value servers, each one a possible target of attack, thus making it difficult to determine the goal of a targeted attacker and to respond appropriately.
  • In an enterprise network, which may include hundreds of thousands of network entities such as laptops, desktop computers, and servers, the network entities can be categorized into different classes. For example, an entity may be a web server, an SQL server, a user terminal, etc. In a strongly connected network, the removal of a small number of connections will not partition the network into isolated parts. At present, however, detection and response systems do not provide adequate insight to system operators as to how best to respond to a strategic attacker. In real-life networks, targets are numerous and easily reachable, making existing approaches that assume a small target set impractical to use.
  • SUMMARY
  • A method for determining cyber-attack targets includes collecting and storing network event information from a plurality of sensors to extract information regarding an attacker; forming an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets; calculating a likelihood for each of the plurality of paths using a processor; and calculating a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker.
  • A method for determining cyber-attack targets includes collecting and storing network event information from a plurality of sensors to extract information regarding an attacker; forming an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets; calculating a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker using a processor; determining a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and removing the determined network graph edge.
  • A method for determining cyber-attack targets includes collecting and storing network event information from a plurality of sensors to extract information regarding an attacker; forming an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets; calculating a likelihood for each of the plurality of paths using a processor; and calculating a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker; calculating a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker; determining a network graph edge to remove which minimizes a defender's expected uncertainty over the potential targets; and removing the determined network graph edge.
  • A system for determining cyber-attack target includes a network monitor module configured to collect network event information from sensors in one or more network nodes; a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a likelihood for each of the plurality of paths, and to calculate a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker.
  • A system for determining cyber-attack targets includes a network monitor module configured to collect network event information from sensors in one or more network nodes; a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker, and to determine a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and a network management module configured to remove the determined network graph edge.
  • A system for determining cyber-attack target includes a network monitor module configured to collect network event information from sensors in one or more network nodes; a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a likelihood for each of the plurality of paths, and to calculate a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker, to calculate a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker, and to determine a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and a network management module configured to remove the determined network graph edge.
  • These and other features and advantages will become apparent from the following detailed description of illustrative embodiments thereof, which is to be read in connection with the accompanying drawings.
  • BRIEF DESCRIPTION OF DRAWINGS
  • The disclosure will provide details in the following description of preferred embodiments with reference to the following figures wherein:
  • FIG. 1 is a diagram of a network graph and an associated attack tree according to the present principles;
  • FIG. 2 is an alternative representation of the mapping between the network graph and the associated attack tree according to the present principles;
  • FIG. 3 is a block/flow diagram of a method for determining an attack target according to the present principles;
  • FIG. 4 is a block/flow diagram of an alternative method for determining an attack target according to the present principles;
  • FIG. 5 is a graph showing a reduction in entropy after removing a network graph edge in accordance with the present principles;
  • FIG. 6 is a diagram of a system for determining an attack target according to the present principles.
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • The present principles employ game theory to predict attacker targets. Using a probabilistic model of attacker behavior, the interactions between a network defender and attacker are modeled, allowing the defender to anticipate future steps of the attack and identify the most likely attack targets based on the observed network events. The present principles use attack scenario trees which represent the possible sequences of high-level attack steps that can be executed at different nodes of the network. This approach differs from the attack-response trees used previously, which represent attack steps within a single network host. Attack scenario trees can be constructed based on past incident reports.
  • The interaction between the defender and the attacker is modeled as a two-player Stackelberg game. The defender can use the model to further decrease uncertainty about attack target predictions by blocking specific network paths (and indirectly any attack steps that traverse those paths) and influencing the attacker to reveal their intentions while conducting the attack. This allows defenders the benefit of proactively blocking future attack steps.
  • Referring now to FIG. 1, an exemplary network graph and associated attack scenario tree are shown. A given network 100 may be represented by a graph of interconnected nodes, with each node being a device or point of vulnerability on the network, and each edge representing a communication link between two nodes. Each node also has one or more associated vulnerabilities, which an attacker may exploit to compromise the security of the node. Combining the vulnerabilities with the edges of the graph provide structure for the attack scenario tree. The nodes of the attack scenario tree represent high-level steps of a potential attack on a defended network.
  • In the present example, an attacker 102 may approach network 100 from one of three externally available systems: web server 104, file server 106, and an email or web client 108. Each of these points of attack has an associated vulnerability. For example, web server 104 may be vulnerable to an SQL injection attack 116, and file server 106 may be vulnerable to a respective file server attack 114. Compromising either of these nodes may give access to active directory (AD) server 110, allowing the attacker 102 to gain access to AD credentials 120. The attacker 102 can then use the AD credentials to install a remote access tool 122 on a target device 112. As an alternative approach, the attacker could stage a phishing attack 118 at a web or email client 108, allowing the attacker 102 to steal a user's credentials 124. Either approach will allow the attacker to obtain access 126 to data or services. It should be noted that any given node may be an AND node, where all of the nodes leading up to it must be reached before accessing the AND node, an OR node where any one input node may be reached, or the nodes may implement any other condition or combination of conditions.
  • The present principles allow a defender to monitor the attack in-progress and provide probabilistic information regarding likely attack paths and targets. By blocking the attacker's access to particular connections, the attack scenario tree can be trimmed and certainty regarding the attacker's goals can be increased.
  • To formalize the model, the network 100 may be represented as a graph G=
    Figure US20130318616A1-20131128-P00001
    V, E
    Figure US20130318616A1-20131128-P00002
    , where the nodes V correspond to the services and machines in the network (for example, a web server 104, an SQL server 109, user machines 112), and the edges E correspond to the connections between them. Each node v∈V belongs to a certain type θ(v), where θ:V→Θ. The node types incorporate the information about node vulnerabilities that can be exploited by the attacker 102. The attack scenario trees are thus constructed of elements of the type set Θ. The set of all attack scenario trees known to the defender is denoted by S={s}.
  • Cyber attacks can be assessed from either the point of view of the defender or from the point of view of the attacker. To the defender, the attacks are paths in the network graph G from the attacker's starting point va∈V to a target t∈V. From the point of view of the attacker, an attack is a path from one of the leaves to the root of an attack scenario tree s∈S. Since s is composed of node types θ(v) and not of specific nodes, each path in an attack scenario tree s can correspond to multiple paths in G as long as the sequence of node types in G matches the types in scenario s. Suppose I(s) is the set of all paths in G corresponding to a scenario s=
    Figure US20130318616A1-20131128-P00001
    θq 1 , . . . , θq k
    Figure US20130318616A1-20131128-P00002
    . Then, for each path I∈I(s), I=
    Figure US20130318616A1-20131128-P00001
    vr 1 , . . . , vr k
    Figure US20130318616A1-20131128-P00002
    , one has θ(vr i )=θq i , i=1, . . . , k. I(s) is the set of instantiations of the attack scenario s.
  • Referring now to FIG. 2, an abstract representation of the mapping between the network 100 and the associated attack tree is shown. Each node is identified by a name and a type. In this case, an attacker starting at node 202 can proceed to one of three nodes 204. The attacker may then take different paths to reach one of the three target nodes 206.
  • It can be assumed that the attacker always has control over the starting node 202 va∈V. As the attacker advances towards one of the target nodes 206 t, the set of active nodes a⊂V which are controlled by the attacker expands until t∈A. This process of attacker's expansion over the nodes in V corresponds to a simultaneous expansion of the set
    Figure US20130318616A1-20131128-P00003
    ∈Θ of active node types in the attack scenario tree until
    Figure US20130318616A1-20131128-P00003
    includes the root of the tree. Inferring a probability distribution over the possible sets A and
    Figure US20130318616A1-20131128-P00003
    helps in predicting the attack targets.
  • As noted above, the interaction between the defender and the attacker is modeled as a two-player Stackelberg game. A Stackelberg two-player game models strategic interaction between two intelligent agents, designated the leader and the follower. Each player has a finite set of actions to choose from. The leader's set of actions is marked as Al, and the follower's set of actions is marked as Af. A pair of actions (al,af) chosen by the players is called the outcome of the game. The players' utilities are functions of the outcome. The leader's utility function is denoted by uf(al,af) and the follower's utility function by ul(al,af). The game proceeds as follows. First, the leader chooses (or commits to) a mixed strategy having a probability distribution over the actions in Al. Then, the follower observes the distribution and chooses a best-response which, generally, can be a probability distribution over Af. There is always a pure-strategy best response for the follower. In other words, one of the optimal best-responses is a degenerate distribution. Moreover, an optimal mixed strategy for the leader can be computed in polynomial time using linear programming techniques.
  • For the purpose of computing an optimal leader's strategy to commit to, the follower's preferences can be represented by choosing an action for each distribution over the leader's actions, instead of using the follower's utility function uf. The follower's preferences are written as a mapping f:σ(Al)→Af, where the operator σ(.) denotes a set of distributions over a given finite set.
  • The present principles provide an extension to the described Stackelberg model in which the follower's preferences are described by a mapping f:σ(Al)→(Af). In other words, if the leader commits to a mixed strategy sl∈σ(Al) then the follower plays a mixed strategy f(sl). This extension is called herein a probabilistic Stackelberg model.
  • A probabilistic Stackelberg model can be used to describe irrational behavior of the follower. The present principles provide a probabilistic Stackelberg model in which the function f is a linear mapping from the vector of probabilities describing the leader's mixed strategy to the vector of probabilities describing the follower's mixed strategy.
  • TABLE 1
    R L
    U
    2, 1 4, 0
    D 1, 0 3, 1
  • Table 1 shows the potential increase in utility that the leader can achieve by considering a probabilistic follower model rather than assuming that the follower is perfectly rational and optimizing a known utility function is shown. In the following two-player normal-form game, the leader is the row player and the follower is the column player. The leader has two strategies U,D, and the follower has two strategies L,R. Each cell in the table shows the leader's and the follower's utility for the corresponding choice of actions. If the follower is perfectly rational and always chooses the action that maximizes its utility, the optimal strategy to commit to for the leader is
  • ( 1 2 - ɛ ) U + ( 1 2 + ɛ ) D .
  • As a result, the follower best-responds with R, and the leader gets a utility of approximately 3.5.
  • Note that the follower best-responds with L if the follower plays U with probability 0.5 or higher, and the follower plays D otherwise. However, if the follower is not perfectly rational, the leader's optimal strategy may be different. For example, consider the case in which the follower is actually playing accordingly to a quantal response model. In a quantal response model, each strategy is played with a positive probability proportional to eλuf. Setting λ=5 and assuming that the leader is playing the Stackelberg strategy
  • 1 2 U + 1 2 D ,
  • the leader can get an even higher utility by deviating to play D more frequently, because the follower will still play R with a relatively high probability after such deviation. This example demonstrates the potential benefits of using a probabilistic follower model, whether derived from the follower's utility function or defined directly as a probability function on the set of the defender's actions.
  • Referring now to FIG. 3, a method for determining probable targets of an attack is shown. It is assumed that the defender has the ability to monitor some or all of the nodes in a network 100 to recognize possible intrusion. Block 302 monitors such network events. The defender's history of observations
    Figure US20130318616A1-20131128-P00004
    =
    Figure US20130318616A1-20131128-P00001
    o1, . . . , ok
    Figure US20130318616A1-20131128-P00002
    indicate possible nodes that became active for the attacker at block 304, so that oiV. Block 306 estimates a probability distribution over the sets of active nodes A and active node types
    Figure US20130318616A1-20131128-P00003
    . Toward this end, block 306 computes heuristical likelihood scores l for each instantiation path π in I(s) as follows
  • l ( π ) = max q 1 < < q m , r 1 < < r m i p ( o q i π r i ) j r 1 , , r m p _ ( π j )
  • Here, p(oq i r i ) is the probability of making an observation oq i when the node πr i is attacked, and p(πj) is the probability of the node πj being attacked without triggering any observations. This likelihood score function can be efficiently computed using the following dynamic programming algorithm:

  • l1:i|
    Figure US20130318616A1-20131128-P00004
    1:j)=max(l1:i−1|
    Figure US20130318616A1-20131128-P00004
    1:j−1)pj|pi),

  • p i)l1:i−1|
    Figure US20130318616A1-20131128-P00004
    1:j),

  • l1:i|
    Figure US20130318616A1-20131128-P00004
    1:j−1))
  • Using l, block 308 computes a probability distribution over the targets t∈V under the assumption that the attacker will follow one of the attack scenario trees in S according to the estimated distribution over the active node types
    Figure US20130318616A1-20131128-P00003
    and the active network nodes A. The distribution over the attack targets is estimated using Monte Carlo simulation:
  • P [ t ] = A l ( A ) P [ t A ]
  • Here, P[t|A] is estimated by simulating the attack steps according to the corresponding attack scenario tree starting with active node types
    Figure US20130318616A1-20131128-P00003
    . The method of FIG. 3 is repeated each time an attacker advances the attack to a new node. On each step of the simulation, the attacker chooses a network node v adjacent to the set of active nodes A such that the type φ(v) matches one of the possible next steps in the current state
    Figure US20130318616A1-20131128-P00003
    of the attack scenario tree. Such attack actions are simulated until the set A expands to include one of the targets t, at which point t is assumed to be the intended target of the attack in the current run of the simulation. The frequency with which target t is attacked in the simulations is the estimate of P[t|A].
  • As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing. Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks. The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
  • As the defender observes network events and collects information about the progress and location of attacker 102, the defender can modify the network 100 in order to force the attacker 102 to reveal the intended target 112 of the attack. In real-world applications, one way for the defender to react to an ongoing cyberattack is to block network connections over a certain port. Translated into the present model, the defender has the ability to block a network edge in response to the observed network events. Since there are likely multiple ways from the attacker's starting node 202 to the any target 206 in a realistic computer network 100, blocking one edge will not prevent the attack. However, the defender may improve the future predictions about the attacker's intended target by carefully choosing a single graph edge to block.
  • Referring now to FIG. 4, a method for computing an optimal defender response is shown. Given the observed network events, collected at block 402, block 404 computes a posterior probability distribution over the attacker's state (A,
    Figure US20130318616A1-20131128-P00003
    ). The optimal edge e∈E to block is the one that minimizes the defender's expected future uncertainty over the attack targets given the history
    Figure US20130318616A1-20131128-P00004
    of the observed network events. In other words, block 406 calculates an edge e=arg mineE[H
    Figure US20130318616A1-20131128-P00004
    ]. Block 408 implements that action by, e.g., disabling the network connection represented by edge e.
  • In one scenario, the attacker has only reached node s 202, and all attack paths going from the bottom of the graph to the top are possible. Additionally, the attacker chooses the next possible node to attack uniformly at random. At first, the probability distribution over the targets is (1/3,1/3,1/3). Note that the distribution will remain the same if the attacker chooses node b as the next node to attack. However, if the attacker chooses node a, then it is certain that t1 is the target of the attack. If the attacker chooses node c, targets t2 and t3 are two equally likely targets.
  • Given a probability distribution p over the targets (e.g., as calculated in FIG. 3), the defender's uncertainty is measured as Shannon's entropy H as follows:
  • H = - t P ( t ) log P ( t )
  • In the above example, the entropy which represents the uncertainty over the attack target, given the history of attacked nodes, is computed as follows:
  • H ( s ) = 3 ( - 1 3 log 1 3 ) = 0.477 H ( sb ) = H ( s ) = 0.477 H ( sa ) = 0 H ( sc ) = 2 ( - 1 2 log 1 2 ) = 0.301
  • Referring now to FIG. 5, a graph showing expected entropy for two scenarios is shown. The vertical axis shows the calculated entropy, while the horizontal axis shows which step of the attack is being considered. The expected entropy is plotted with a solid line. The entropy can be reduced, however, if the possibility of the attacker choosing node b is removed. Removing the edge between s and b results in a reduction of the expected entropy after the next attack step. This reduced entropy with a dashed line on the same plot. By step 2, entropy is zero for both plots because the attacker has reached the target t.
  • Using the present principles provides a significant improvement in terms of the defender's uncertainty as to the attacker's potential targets. Moreover, by blocking paths in concordance with the probabilistic Stackelberg model of the present principles, the defender is able to further reduce the uncertainty by influencing the attacker's next move. Furthermore, the present principles are scalable, with good performance in attack paths of length 3, 4, and 5—it should be noted that practical attack paths rarely exceed four steps in length. Simulations run with over 1000 nodes and with attack path lengths of five were able to complete in a matter of seconds, making the present principles practical for real-world application.
  • Referring now to FIG. 6, a system for network defense is shown. A defender employs a defender administration terminal 602 to observe and respond to an attacker's actions. The defender administration terminal 602 may be a stand-alone device, or may be part of a more general purpose administration terminal. The defender terminal 602 communicates with sensors 604 distributed throughout a network 100 to collect information regarding potential intrusions. The sensors may be incorporated in any node in a network 100 and may monitor any process or component to detect irregular activities. Multiple sensors 604 may be incorporated in a single node. Sensors 604 may provide detailed information regarding the attacker and its activities, may report on any irregular activity, or may simply represent logs for all activities in the system, said logs to be analyzed by the defender terminal 602. The sensors 604 may include software installed on nodes to monitor system calls and user activities, or alternatively be standalone devices such as traffic sniffers configured to detect particular suspicious forms of network traffic to or from monitored nodes.
  • The defender administration terminal 602 processes the data provided by sensors 604 and determines likely targets for the attacker. The defender terminal 602 also calculates an optimal defender response to reduce the uncertainty in the attacker's targets. Defender terminal 602 includes a processor 608 and memory 610 to collect and utilize the sensor data using network monitor 614. The network monitor 614 collects data from sensors 604 and processes that data from potentially heterogeneous sources into a usable form. The network monitor may, for example, parse logs provided by sensors 604 to find suspicious or abnormal entries. The processor 608 uses the sensor data provided by network monitor 614 and stored in memory 610 to produce the most probable targets and response.
  • Once the processor 608 calculates an optimal response, the network control module 612 executes the response using a network management interface 606. The network management 606 interface may represent any appropriate form of network management, including for example a simple network management protocol (SNMP) device. In this manner, the defender administration terminal 602 can disconnect links in a network 100, or take similar network-level measures that prevent the attacker to proceed along the chosen network links (e.g., the defender could choose to enable a firewall system on that link, instead of disconnecting the link). The changed network topology forces the attacker along different paths, and the attacker's response to the defender's action substantially reduces the uncertainty regarding the attacker's intentions.
  • Having described preferred embodiments of a system and method for predicting attacks based on probabilistic game-theory (which are intended to be illustrative and not limiting), it is noted that modifications and variations can be made by persons skilled in the art in light of the above teachings. It is therefore to be understood that changes may be made in the particular embodiments disclosed which are within the scope of the invention as outlined by the appended claims. Having thus described aspects of the invention, with the details and particularity required by the patent laws, what is claimed and desired protected by Letters Patent is set forth in the appended claims.

Claims (20)

1. A system for determining cyber-attack target, comprising:
a network monitor module configured to collect network event information from sensors in one or more network nodes;
a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a likelihood for each of the plurality of paths, and to calculate a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker.
2. The system of claim 1, wherein the network event information comprises network traffic information.
3. The system of claim 1, wherein the network event information comprises node system calls.
4. The system of claim 1, wherein the network event information comprises activity logs.
5. The system of claim 1, wherein the processor is configured to compute likelihood l for each of the plurality of paths π as
l ( π ) = max q 1 < < q m , r 1 < < r m i p ( o q i π r i ) j r 1 , , r m p _ ( π j ) ,
where p(oq i r i ) is a probability of making an observation oq i when a node πr i is attacked and pj) is the probability of the node πj being attacked without triggering any observations.
6. The system of claim 1, wherein processor is configured to compute the probability distribution for the set of potential targets using a Monte Carlo simulation.
7. The system of claim 1, wherein the processor is configured to compute the probability distribution for the set of potential targets as
P [ t ] = A l ( A ) P [ t A ] ,
where P]t|A] is a probability of an attacker pursuing a target node t given a set of occupied nodes A, and l(A) is a likelihood of the set of nodes A being occupied.
8. A system for determining cyber-attack targets, comprising:
a network monitor module configured to collect network event information from sensors in one or more network nodes;
a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker, and to determine a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and
a network management module configured to remove the determined network graph edge.
9. The system of claim 8, wherein the network event information comprises network traffic information.
10. The system of claim 8, wherein the network event information comprises node system calls.
11. The system of claim 8, wherein the network event information comprises activity logs.
12. The system of claim 8, wherein the processor is configured to determine the defender's expected uncertainty over the potential targets as a Shannon entropy.
13. The system of claim 8, wherein the network management module is configured to remove the determined network graph edge by disconnecting a corresponding network communication link.
14. The system of claim 8, wherein the network management module is configured to remove the determined network graph edge by adapting a corresponding network communication link to block the attacker's use of the determined network graph edge.
15. A system for determining cyber-attack target, comprising:
a network monitor module configured to collect network event information from sensors in one or more network nodes;
a processor configured to extract information regarding an attacker from the network event information, to form an attack scenario tree that encodes network topology and vulnerability information including a plurality of paths from known compromised nodes to a set of potential targets, to calculate a likelihood for each of the plurality of paths, to calculate a probability distribution for the set of potential targets to determine which potential targets are most likely pursued by the attacker, to calculate a probability distribution over a set of nodes and node vulnerability types already accessed by the attacker, and to determine a network graph edge to remove that minimizes a defender's expected uncertainty over the potential targets; and
a network management module configured to remove the determined network graph edge.
16. The system of claim 15, wherein the network event information comprises network traffic information.
17. The system of claim 15, wherein the network event information comprises node system calls.
18. The system of claim 15, wherein the network event information comprises activity logs.
19. The system of claim 15, wherein the network management module is configured to remove the determined network graph edge by disconnecting a corresponding network communication link.
20. The system of claim 15, wherein the network management module is configured to remove the determined network graph edge by adapting a corresponding network communication link to block the attacker's use of the determined network graph edge.
US13/487,774 2012-05-23 2012-06-04 Predicting attacks based on probabilistic game-theory Abandoned US20130318616A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/487,774 US20130318616A1 (en) 2012-05-23 2012-06-04 Predicting attacks based on probabilistic game-theory

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/478,290 US8863293B2 (en) 2012-05-23 2012-05-23 Predicting attacks based on probabilistic game-theory
US13/487,774 US20130318616A1 (en) 2012-05-23 2012-06-04 Predicting attacks based on probabilistic game-theory

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/478,290 Continuation US8863293B2 (en) 2012-05-23 2012-05-23 Predicting attacks based on probabilistic game-theory

Publications (1)

Publication Number Publication Date
US20130318616A1 true US20130318616A1 (en) 2013-11-28

Family

ID=49622628

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/478,290 Active 2032-07-28 US8863293B2 (en) 2012-05-23 2012-05-23 Predicting attacks based on probabilistic game-theory
US13/487,774 Abandoned US20130318616A1 (en) 2012-05-23 2012-06-04 Predicting attacks based on probabilistic game-theory

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/478,290 Active 2032-07-28 US8863293B2 (en) 2012-05-23 2012-05-23 Predicting attacks based on probabilistic game-theory

Country Status (1)

Country Link
US (2) US8863293B2 (en)

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130273514A1 (en) * 2007-10-15 2013-10-17 University Of Southern California Optimal Strategies in Security Games
CN104883345A (en) * 2014-09-28 2015-09-02 宁波匡恩网络科技有限公司 Network security feature automatic deployment method and system
US9141790B2 (en) * 2013-09-10 2015-09-22 Symantec Corporation Systems and methods for using event-correlation graphs to detect attacks on computing systems
US9166997B1 (en) * 2013-09-19 2015-10-20 Symantec Corporation Systems and methods for reducing false positives when using event-correlation graphs to detect attacks on computing systems
US20160028758A1 (en) * 2014-03-28 2016-01-28 Zitovault, Inc. System and Method for Predicting Impending Cyber Security Events Using Multi Channel Behavioral Analysis in a Distributed Computing Environment
US9369484B1 (en) * 2014-07-24 2016-06-14 Rockwell Collins, Inc. Dynamic security hardening of security critical functions
US20160226893A1 (en) * 2015-01-30 2016-08-04 Wipro Limited Methods for optimizing an automated determination in real-time of a risk rating of cyber-attack and devices thereof
US20160239787A1 (en) * 2015-02-12 2016-08-18 Fujitsu Limited Security plan support method, security plan support device and recording medium
CN107193658A (en) * 2017-05-25 2017-09-22 重庆工程学院 Cloud computing resource scheduling method based on game theory
US20170279820A1 (en) * 2016-03-24 2017-09-28 Charles Dale Herring System and method for detecting computer attacks
CN107566387A (en) * 2017-09-14 2018-01-09 中国人民解放军信息工程大学 Cyber-defence action decision method based on attacking and defending evolutionary Game Analysis
US10007792B1 (en) * 2013-11-27 2018-06-26 Amazon Technologies, Inc. Modeling a network security environment as a game
US10027698B2 (en) * 2015-12-18 2018-07-17 National Technology & Engineering Solutions Of Sandia, Llc Network protection system using linkographs
US20180307832A1 (en) * 2015-12-14 2018-10-25 Mitsubishi Electric Corporation Information processing device, information processing method, and computer readable medium
US10182065B1 (en) 2016-06-29 2019-01-15 Rockwell Collins, Inc. Distributed vetronics intrustion detection system and device
US10425429B2 (en) * 2013-04-10 2019-09-24 Gabriel Bassett System and method for cyber security analysis and human behavior prediction
US10430581B2 (en) * 2016-12-22 2019-10-01 Chronicle Llc Computer telemetry analysis
US10440036B2 (en) * 2015-12-09 2019-10-08 Checkpoint Software Technologies Ltd Method and system for modeling all operations and executions of an attack and malicious process entry
CN110398981A (en) * 2019-06-20 2019-11-01 合肥工业大学 Strategy space tailoring method and system for UAV formation task assignment
CN110417733A (en) * 2019-06-24 2019-11-05 中国人民解放军战略支援部队信息工程大学 Attack prediction method, device and system based on QBD attack and defense random evolutionary game model
CN111277561A (en) * 2019-12-27 2020-06-12 北京威努特技术有限公司 Network attack path prediction method and device and security management platform
US20200201989A1 (en) * 2018-10-12 2020-06-25 International Business Machines Corporation Multi-point causality tracking in cyber incident reasoning
US10880316B2 (en) 2015-12-09 2020-12-29 Check Point Software Technologies Ltd. Method and system for determining initial execution of an attack
US20210064762A1 (en) * 2019-08-29 2021-03-04 Darktrace Limited Intelligent adversary simulator
US20210194924A1 (en) * 2019-08-29 2021-06-24 Darktrace Limited Artificial intelligence adversary red team
US20210288992A1 (en) * 2020-03-16 2021-09-16 Otorio Ltd. Operational Network Risk Mitigation System And Method
US20210357501A1 (en) * 2019-03-12 2021-11-18 Mitsubishi Electric Corporation Attack estimation device, attack estimation method, and attack estimation program
US11228517B2 (en) * 2015-03-06 2022-01-18 Georgia Tech Research Corporation Device fingerprinting for cyber-physical systems
CN113992355A (en) * 2021-09-28 2022-01-28 新华三信息安全技术有限公司 Attack prediction method, device, equipment and machine readable storage medium
US20220053015A1 (en) * 2015-10-12 2022-02-17 Palantir Technologies Inc. Systems for computer network security risk assessment including user compromise analysis associated with a network of devices
US20220060509A1 (en) * 2015-10-28 2022-02-24 Qomplx, Inc. Privilege assurance of enterprise computer network environments using lateral movement detection and prevention
US20220060507A1 (en) * 2015-10-28 2022-02-24 Qomplx, Inc. Privilege assurance of enterprise computer network environments using attack path detection and prediction
US20220229903A1 (en) * 2021-01-21 2022-07-21 Intuit Inc. Feature extraction and time series anomaly detection over dynamic graphs
CN114844668A (en) * 2022-03-17 2022-08-02 清华大学 Defense resource configuration method, device, equipment and readable medium
US20220360597A1 (en) * 2019-08-29 2022-11-10 Darktrace Holdings Limited Cyber security system utilizing interactions between detected and hypothesize cyber-incidents
US20230275913A1 (en) * 2022-02-25 2023-08-31 Microsoft Technology Licensing, Llc Using graph enrichment to detect a potentially malicious access attempt
CN117499158A (en) * 2023-12-25 2024-02-02 天地信息网络研究院(安徽)有限公司 Active defense method based on multi-attacker joint or non-joint attack
CN120090865A (en) * 2025-04-25 2025-06-03 证通股份有限公司 Method, device, medium and program product for determining attack and defense information in a heterogeneous network

Families Citing this family (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2013272211B2 (en) 2012-03-22 2016-11-24 Los Alamos National Security, Llc Path scanning for the detection of anomalous subgraphs, anomaly/change detection and network situational awareness
US20140274246A1 (en) * 2013-03-15 2014-09-18 University Of Southern California Localized shortest-paths estimation of influence propagation for multiple influencers
US9544380B2 (en) * 2013-04-10 2017-01-10 International Business Machines Corporation Data analytics and security in social networks
US10574675B2 (en) * 2014-12-05 2020-02-25 T-Mobile Usa, Inc. Similarity search for discovering multiple vector attacks
US10216938B2 (en) 2014-12-05 2019-02-26 T-Mobile Usa, Inc. Recombinant threat modeling
AU2016204072B2 (en) 2015-06-17 2017-08-03 Accenture Global Services Limited Event anomaly analysis and prediction
US10043018B2 (en) 2015-11-17 2018-08-07 Microsoft Technology Licensing, Llc Access privilege analysis for a securable asset
US10313385B2 (en) 2015-11-30 2019-06-04 Arizona Board Of Regents On Behalf Of Arizona State University Systems and methods for data driven game theoretic cyber threat mitigation
EP3371947B1 (en) * 2015-12-14 2020-09-02 Siemens Aktiengesellschaft System and method for passive assessment of industrial perimeter security
US10264009B2 (en) * 2016-07-26 2019-04-16 Booz Allen Hamilton Inc. Automated machine learning scheme for software exploit prediction
US10796284B2 (en) 2016-09-20 2020-10-06 Fujitsu Limited Collaborative scheduling
GB201622449D0 (en) * 2016-12-30 2017-02-15 Bt Group Legal Ip Dept Identifying an attacked computing device
US10771483B2 (en) * 2016-12-30 2020-09-08 British Telecommunications Public Limited Company Identifying an attacked computing device
FR3063554B1 (en) * 2017-03-03 2021-04-02 Mbda France METHOD AND DEVICE FOR PREDICTING OPTIMAL ATTACK AND DEFENSE SOLUTIONS IN A MILITARY CONFLICT SCENARIO
CN110999249A (en) * 2017-08-03 2020-04-10 T移动美国公司 Similarity search for discovering multiple vector attacks
CN107888588B (en) * 2017-11-09 2020-07-24 上海海事大学 A K-Maximum Probability Attack Path Solving Method for Specified Target Node Sets
CN107733917B (en) * 2017-11-09 2020-07-24 上海海事大学 A Progressive Solution Method for K-Maximum Probability Attack Paths of Specified Target Node Sets
US11010233B1 (en) 2018-01-18 2021-05-18 Pure Storage, Inc Hardware-based system monitoring
CN109617863B (en) * 2018-11-27 2020-02-18 杭州电子科技大学 A Game Theory-Based Approach to Optimal Defense Strategy Selection for Moving Target Defense
CN110401650A (en) * 2019-07-17 2019-11-01 五邑大学 A network security decision-making method, device and storage medium for game attack and defense graph
CN110784487B (en) * 2019-11-07 2021-08-31 广东技术师范大学 A SDN node defense method based on data packet sampling model
US12050689B2 (en) 2019-11-22 2024-07-30 Pure Storage, Inc. Host anomaly-based generation of snapshots
US11755751B2 (en) * 2019-11-22 2023-09-12 Pure Storage, Inc. Modify access restrictions in response to a possible attack against data stored by a storage system
US12079333B2 (en) * 2019-11-22 2024-09-03 Pure Storage, Inc. Independent security threat detection and remediation by storage systems in a synchronous replication arrangement
US12204657B2 (en) 2019-11-22 2025-01-21 Pure Storage, Inc. Similar block detection-based detection of a ransomware attack
US12067118B2 (en) * 2019-11-22 2024-08-20 Pure Storage, Inc. Detection of writing to a non-header portion of a file as an indicator of a possible ransomware attack against a storage system
US11675898B2 (en) 2019-11-22 2023-06-13 Pure Storage, Inc. Recovery dataset management for security threat monitoring
US11720692B2 (en) 2019-11-22 2023-08-08 Pure Storage, Inc. Hardware token based management of recovery datasets for a storage system
US11687418B2 (en) 2019-11-22 2023-06-27 Pure Storage, Inc. Automatic generation of recovery plans specific to individual storage elements
US12079356B2 (en) 2019-11-22 2024-09-03 Pure Storage, Inc. Measurement interval anomaly detection-based generation of snapshots
US11651075B2 (en) * 2019-11-22 2023-05-16 Pure Storage, Inc. Extensible attack monitoring by a storage system
US12050683B2 (en) 2019-11-22 2024-07-30 Pure Storage, Inc. Selective control of a data synchronization setting of a storage system based on a possible ransomware attack against the storage system
US12079502B2 (en) * 2019-11-22 2024-09-03 Pure Storage, Inc. Storage element attribute-based determination of a data protection policy for use within a storage system
US11720714B2 (en) 2019-11-22 2023-08-08 Pure Storage, Inc. Inter-I/O relationship based detection of a security threat to a storage system
US12153670B2 (en) 2019-11-22 2024-11-26 Pure Storage, Inc. Host-driven threat detection-based protection of storage elements within a storage system
US11625481B2 (en) * 2019-11-22 2023-04-11 Pure Storage, Inc. Selective throttling of operations potentially related to a security threat to a storage system
US11341236B2 (en) 2019-11-22 2022-05-24 Pure Storage, Inc. Traffic-based detection of a security threat to a storage system
US11941116B2 (en) 2019-11-22 2024-03-26 Pure Storage, Inc. Ransomware-based data protection parameter modification
US20210382992A1 (en) * 2019-11-22 2021-12-09 Pure Storage, Inc. Remote Analysis of Potentially Corrupt Data Written to a Storage System
US12248566B2 (en) * 2019-11-22 2025-03-11 Pure Storage, Inc. Snapshot deletion pattern-based determination of ransomware attack against data maintained by a storage system
US11520907B1 (en) * 2019-11-22 2022-12-06 Pure Storage, Inc. Storage system snapshot retention based on encrypted data
US11645162B2 (en) 2019-11-22 2023-05-09 Pure Storage, Inc. Recovery point determination for data restoration in a storage system
US11657155B2 (en) 2019-11-22 2023-05-23 Pure Storage, Inc Snapshot delta metric based determination of a possible ransomware attack against data maintained by a storage system
US12411962B2 (en) * 2019-11-22 2025-09-09 Pure Storage, Inc. Managed run-time environment-based detection of a ransomware attack
US11483319B2 (en) 2020-03-05 2022-10-25 International Business Machines Corporation Security model
US12093396B2 (en) 2020-07-16 2024-09-17 Bank Of America Corporation System and method for associating a common vulnerability and exposures (CVE) with a computing device and applying a security patch
CN112712131B (en) * 2021-01-14 2025-03-07 清华大学 Lifelong learning method of neural network model based on game theory framework
CN112948881A (en) * 2021-03-16 2021-06-11 西安电子科技大学 Method for calculating information leakage probability in open scene
CN113312625B (en) * 2021-06-21 2024-01-02 深信服科技股份有限公司 Attack path diagram construction method, device, equipment and medium
US12294582B2 (en) 2022-06-03 2025-05-06 Kbr Wyle Services, Llc Communication network resilience based on distributed multi-agent q-learning algorithm
CN115550031B (en) * 2022-09-28 2025-04-25 中国农业银行股份有限公司 Network defense method, control device and storage medium
CN116886379B (en) * 2023-07-21 2024-05-14 鹏城实验室 Network attack reconstruction method, model training method and related devices
CN118555599B (en) * 2024-07-26 2024-10-01 中国人民解放军国防科技大学 Multi-objective optimization-based wireless sensor network key node identification method
CN119420569B (en) * 2024-11-29 2025-09-23 中国联合网络通信集团有限公司 Attack tracing path determining method, device, equipment, storage medium and product

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020059078A1 (en) * 2000-09-01 2002-05-16 Valdes Alfonso De Jesus Probabilistic alert correlation
US7162741B2 (en) * 2001-07-30 2007-01-09 The Trustees Of Columbia University In The City Of New York System and methods for intrusion detection with dynamic window sizes
US20090287623A1 (en) * 2006-09-21 2009-11-19 Rutgers, The State University Of New Jersey Method and apparatus for using bayesian networks for localization
US20090307772A1 (en) * 2008-05-21 2009-12-10 Honeywell International Inc. framework for scalable state estimation using multi network observations
US20110061104A1 (en) * 2009-09-08 2011-03-10 Core Sdi, Inc. System and method for probabilistic attack planning
US20110288692A1 (en) * 2010-05-20 2011-11-24 Accenture Global Services Gmbh Malicious attack detection and analysis
US20120180126A1 (en) * 2010-07-13 2012-07-12 Lei Liu Probable Computing Attack Detector
US8549641B2 (en) * 2009-09-03 2013-10-01 Palo Alto Research Center Incorporated Pattern-based application classification

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7013395B1 (en) * 2001-03-13 2006-03-14 Sandra Corporation Method and tool for network vulnerability analysis
US8359650B2 (en) * 2002-10-01 2013-01-22 Skybox Secutiry Inc. System, method and computer readable medium for evaluating potential attacks of worms
US8046835B2 (en) 2002-10-23 2011-10-25 Frederick S. M. Herz Distributed computer network security activity model SDI-SCAM
US7685085B2 (en) * 2003-11-10 2010-03-23 James Ralph Heidenreich System and method to facilitate user thinking about an arbitrary problem with output and interfaces to external systems, components and resources
JP4371905B2 (en) 2004-05-27 2009-11-25 富士通株式会社 Unauthorized access detection device, unauthorized access detection method, unauthorized access detection program, and distributed service disablement attack detection device
US7590589B2 (en) 2004-09-10 2009-09-15 Hoffberg Steven M Game theoretic prioritization scheme for mobile ad hoc networks permitting hierarchal deference
US7530105B2 (en) * 2006-03-21 2009-05-05 21St Century Technologies, Inc. Tactical and strategic attack detection and prediction
US9171157B2 (en) * 2006-03-28 2015-10-27 Blue Coat Systems, Inc. Method and system for tracking access to application data and preventing data exploitation by malicious programs
US8539586B2 (en) * 2006-05-19 2013-09-17 Peter R. Stephenson Method for evaluating system risk
US7971252B2 (en) * 2006-06-09 2011-06-28 Massachusetts Institute Of Technology Generating a multiple-prerequisite attack graph
US20080109272A1 (en) * 2006-11-08 2008-05-08 Anshul Sheopuri Apparatus, System, Method and Computer Program Product for Analysis of Fraud in Transaction Data
CN101075917B (en) * 2007-07-16 2010-08-25 华为技术有限公司 Method and apparatus for predicting network attack behaviour
US7813739B2 (en) * 2007-09-27 2010-10-12 Koon Hoo Teo Method for reducing inter-cell interference in wireless OFDMA networks
US8195490B2 (en) * 2007-10-15 2012-06-05 University Of Southern California Agent security via approximate solvers
US8224681B2 (en) * 2007-10-15 2012-07-17 University Of Southern California Optimizing a security patrolling strategy using decomposed optimal Bayesian Stackelberg solver
KR101445188B1 (en) * 2007-11-21 2014-10-01 삼성전자주식회사 Abnormal frame detection method in wireless sensor network and system therefor
US20100058456A1 (en) 2008-08-27 2010-03-04 Sushil Jajodia IDS Sensor Placement Using Attack Graphs
US8108188B2 (en) * 2008-10-30 2012-01-31 Honeywell International Inc. Enumerated linear programming for optimal strategies
US7961137B2 (en) 2008-11-10 2011-06-14 The Boeing Company System and method for detecting performance of a sensor field at all points within a geographic area of regard
US8671277B2 (en) 2009-06-09 2014-03-11 Syracuse University Method for secure communication over heterogeneous networks
GB201004226D0 (en) 2010-03-15 2010-04-28 Bae Systems Plc Process analysis
WO2011162848A2 (en) 2010-04-01 2011-12-29 21Ct, Inc. System and method for providing impact modeling and prediction of attacks on cyber targets
US9201131B2 (en) 2010-11-18 2015-12-01 The Boeing Company Secure routing based on degree of trust
US9515826B2 (en) 2010-11-18 2016-12-06 The Boeing Company Network topology aided by smart agent download

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020059078A1 (en) * 2000-09-01 2002-05-16 Valdes Alfonso De Jesus Probabilistic alert correlation
US7162741B2 (en) * 2001-07-30 2007-01-09 The Trustees Of Columbia University In The City Of New York System and methods for intrusion detection with dynamic window sizes
US20090287623A1 (en) * 2006-09-21 2009-11-19 Rutgers, The State University Of New Jersey Method and apparatus for using bayesian networks for localization
US20090307772A1 (en) * 2008-05-21 2009-12-10 Honeywell International Inc. framework for scalable state estimation using multi network observations
US8549641B2 (en) * 2009-09-03 2013-10-01 Palo Alto Research Center Incorporated Pattern-based application classification
US20110061104A1 (en) * 2009-09-08 2011-03-10 Core Sdi, Inc. System and method for probabilistic attack planning
US20110288692A1 (en) * 2010-05-20 2011-11-24 Accenture Global Services Gmbh Malicious attack detection and analysis
US20120180126A1 (en) * 2010-07-13 2012-07-12 Lei Liu Probable Computing Attack Detector

Cited By (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130273514A1 (en) * 2007-10-15 2013-10-17 University Of Southern California Optimal Strategies in Security Games
US10425429B2 (en) * 2013-04-10 2019-09-24 Gabriel Bassett System and method for cyber security analysis and human behavior prediction
US9141790B2 (en) * 2013-09-10 2015-09-22 Symantec Corporation Systems and methods for using event-correlation graphs to detect attacks on computing systems
US9166997B1 (en) * 2013-09-19 2015-10-20 Symantec Corporation Systems and methods for reducing false positives when using event-correlation graphs to detect attacks on computing systems
US10007792B1 (en) * 2013-11-27 2018-06-26 Amazon Technologies, Inc. Modeling a network security environment as a game
US20160028758A1 (en) * 2014-03-28 2016-01-28 Zitovault, Inc. System and Method for Predicting Impending Cyber Security Events Using Multi Channel Behavioral Analysis in a Distributed Computing Environment
US9602530B2 (en) * 2014-03-28 2017-03-21 Zitovault, Inc. System and method for predicting impending cyber security events using multi channel behavioral analysis in a distributed computing environment
US9369484B1 (en) * 2014-07-24 2016-06-14 Rockwell Collins, Inc. Dynamic security hardening of security critical functions
CN104883345A (en) * 2014-09-28 2015-09-02 宁波匡恩网络科技有限公司 Network security feature automatic deployment method and system
US20160226893A1 (en) * 2015-01-30 2016-08-04 Wipro Limited Methods for optimizing an automated determination in real-time of a risk rating of cyber-attack and devices thereof
US10546256B2 (en) * 2015-02-12 2020-01-28 Fujitsu Limited Security plan support method, security plan support device and recording medium
US20160239787A1 (en) * 2015-02-12 2016-08-18 Fujitsu Limited Security plan support method, security plan support device and recording medium
US11228517B2 (en) * 2015-03-06 2022-01-18 Georgia Tech Research Corporation Device fingerprinting for cyber-physical systems
US20220053015A1 (en) * 2015-10-12 2022-02-17 Palantir Technologies Inc. Systems for computer network security risk assessment including user compromise analysis associated with a network of devices
US11956267B2 (en) * 2015-10-12 2024-04-09 Palantir Technologies Inc. Systems for computer network security risk assessment including user compromise analysis associated with a network of devices
US12113831B2 (en) * 2015-10-28 2024-10-08 Qomplx Llc Privilege assurance of enterprise computer network environments using lateral movement detection and prevention
US20220060509A1 (en) * 2015-10-28 2022-02-24 Qomplx, Inc. Privilege assurance of enterprise computer network environments using lateral movement detection and prevention
US20220060507A1 (en) * 2015-10-28 2022-02-24 Qomplx, Inc. Privilege assurance of enterprise computer network environments using attack path detection and prediction
US12107895B2 (en) * 2015-10-28 2024-10-01 Qomplx Llc Privilege assurance of enterprise computer network environments using attack path detection and prediction
US10880316B2 (en) 2015-12-09 2020-12-29 Check Point Software Technologies Ltd. Method and system for determining initial execution of an attack
US10440036B2 (en) * 2015-12-09 2019-10-08 Checkpoint Software Technologies Ltd Method and system for modeling all operations and executions of an attack and malicious process entry
US20200084230A1 (en) * 2015-12-09 2020-03-12 Check Point Software Technologies Ltd. Method And System For Modeling All Operations And Executions Of An Attack And Malicious Process Entry
US10972488B2 (en) * 2015-12-09 2021-04-06 Check Point Software Technologies Ltd. Method and system for modeling all operations and executions of an attack and malicious process entry
US20180307832A1 (en) * 2015-12-14 2018-10-25 Mitsubishi Electric Corporation Information processing device, information processing method, and computer readable medium
US10027698B2 (en) * 2015-12-18 2018-07-17 National Technology & Engineering Solutions Of Sandia, Llc Network protection system using linkographs
US20170279820A1 (en) * 2016-03-24 2017-09-28 Charles Dale Herring System and method for detecting computer attacks
US10182065B1 (en) 2016-06-29 2019-01-15 Rockwell Collins, Inc. Distributed vetronics intrustion detection system and device
US10430581B2 (en) * 2016-12-22 2019-10-01 Chronicle Llc Computer telemetry analysis
US10839071B2 (en) * 2016-12-22 2020-11-17 Chronicle Llc Computer telemetry analysis
CN107193658A (en) * 2017-05-25 2017-09-22 重庆工程学院 Cloud computing resource scheduling method based on game theory
CN107566387A (en) * 2017-09-14 2018-01-09 中国人民解放军信息工程大学 Cyber-defence action decision method based on attacking and defending evolutionary Game Analysis
US10956566B2 (en) * 2018-10-12 2021-03-23 International Business Machines Corporation Multi-point causality tracking in cyber incident reasoning
US20200201989A1 (en) * 2018-10-12 2020-06-25 International Business Machines Corporation Multi-point causality tracking in cyber incident reasoning
US20210357501A1 (en) * 2019-03-12 2021-11-18 Mitsubishi Electric Corporation Attack estimation device, attack estimation method, and attack estimation program
US11893110B2 (en) * 2019-03-12 2024-02-06 Mitsubishi Electric Corporation Attack estimation device, attack estimation method, and attack estimation program
CN110398981A (en) * 2019-06-20 2019-11-01 合肥工业大学 Strategy space tailoring method and system for UAV formation task assignment
CN110417733A (en) * 2019-06-24 2019-11-05 中国人民解放军战略支援部队信息工程大学 Attack prediction method, device and system based on QBD attack and defense random evolutionary game model
US20240333763A1 (en) * 2019-08-29 2024-10-03 Darktrace Holdings Limited Artificial intelligence adversary red team
US20210064762A1 (en) * 2019-08-29 2021-03-04 Darktrace Limited Intelligent adversary simulator
US12326943B2 (en) * 2019-08-29 2025-06-10 Darktrace Holdings Limited Intelligent adversary simulator
US20220360597A1 (en) * 2019-08-29 2022-11-10 Darktrace Holdings Limited Cyber security system utilizing interactions between detected and hypothesize cyber-incidents
US20210194924A1 (en) * 2019-08-29 2021-06-24 Darktrace Limited Artificial intelligence adversary red team
US11709944B2 (en) * 2019-08-29 2023-07-25 Darktrace Holdings Limited Intelligent adversary simulator
US12034767B2 (en) * 2019-08-29 2024-07-09 Darktrace Holdings Limited Artificial intelligence adversary red team
US20230351027A1 (en) * 2019-08-29 2023-11-02 Darktrace Holdings Limited Intelligent adversary simulator
CN111277561A (en) * 2019-12-27 2020-06-12 北京威努特技术有限公司 Network attack path prediction method and device and security management platform
US11637853B2 (en) * 2020-03-16 2023-04-25 Otorio Ltd. Operational network risk mitigation system and method
US20210288992A1 (en) * 2020-03-16 2021-09-16 Otorio Ltd. Operational Network Risk Mitigation System And Method
US20220229903A1 (en) * 2021-01-21 2022-07-21 Intuit Inc. Feature extraction and time series anomaly detection over dynamic graphs
US12118077B2 (en) * 2021-01-21 2024-10-15 Intuit Inc. Feature extraction and time series anomaly detection over dynamic graphs
CN113992355A (en) * 2021-09-28 2022-01-28 新华三信息安全技术有限公司 Attack prediction method, device, equipment and machine readable storage medium
US20230275913A1 (en) * 2022-02-25 2023-08-31 Microsoft Technology Licensing, Llc Using graph enrichment to detect a potentially malicious access attempt
US12231448B2 (en) * 2022-02-25 2025-02-18 Microsoft Technology Licensing, Llc Using graph enrichment to detect a potentially malicious access attempt
CN114844668A (en) * 2022-03-17 2022-08-02 清华大学 Defense resource configuration method, device, equipment and readable medium
CN117499158A (en) * 2023-12-25 2024-02-02 天地信息网络研究院(安徽)有限公司 Active defense method based on multi-attacker joint or non-joint attack
CN120090865A (en) * 2025-04-25 2025-06-03 证通股份有限公司 Method, device, medium and program product for determining attack and defense information in a heterogeneous network

Also Published As

Publication number Publication date
US8863293B2 (en) 2014-10-14
US20130318615A1 (en) 2013-11-28

Similar Documents

Publication Publication Date Title
US8863293B2 (en) Predicting attacks based on probabilistic game-theory
US12003534B2 (en) Detecting and mitigating forged authentication attacks within a domain
US11968227B2 (en) Detecting KERBEROS ticket attacks within a domain
US20220201042A1 (en) Ai-driven defensive penetration test analysis and recommendation system
US11582207B2 (en) Detecting and mitigating forged authentication object attacks using an advanced cyber decision platform
US11757849B2 (en) Detecting and mitigating forged authentication object attacks in multi-cloud environments
US20230118726A1 (en) System and methods for detecting and mitigating golden saml attacks against federated services
US10560483B2 (en) Rating organization cybersecurity using active and passive external reconnaissance
US20230370490A1 (en) System and method for cyber exploitation path analysis and task plan optimization
JP6201614B2 (en) Log analysis apparatus, method and program
US20140157415A1 (en) Information security analysis using game theory and simulation
US20230388278A1 (en) Detecting and mitigating forged authentication object attacks in multi - cloud environments with attestation
JP7019533B2 (en) Attack detection device, attack detection system, attack detection method and attack detection program
Xie et al. An adaptive defense mechanism to prevent advanced persistent threats
CN113518086B (en) Network attack prediction method, device and storage medium
CN114448718A (en) A network security assurance method for parallel detection and repair
Hewett et al. Smart Grid security: Deriving informed decisions from cyber attack game analysis
Luo et al. A fictitious play‐based response strategy for multistage intrusion defense systems
Paulauskas et al. Attacker skill level distribution estimation in the system mean time-to-compromise
WO2019113492A1 (en) Detecting and mitigating forged authentication object attacks using an advanced cyber decision platform
Al Abri et al. Game-theoretic modeling of the behavior of Domain Name System attacker
Meetei Mathematical model of security approaches on cloud computing
Abazari et al. Optimal response to computer network threats
Jia et al. A Deception Defense Strategy Based on Game Theory
Fu et al. A study of evaluation methods of WEB security threats based on multi-stage attack

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION