[go: up one dir, main page]

GB2568114A - Dynamic security policy - Google Patents

Dynamic security policy Download PDF

Info

Publication number
GB2568114A
GB2568114A GB1718427.6A GB201718427A GB2568114A GB 2568114 A GB2568114 A GB 2568114A GB 201718427 A GB201718427 A GB 201718427A GB 2568114 A GB2568114 A GB 2568114A
Authority
GB
United Kingdom
Prior art keywords
configuration
vector
vulnerability
training
security
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB1718427.6A
Other versions
GB2568114B (en
GB201718427D0 (en
Inventor
Shackleton Mark
El-Moussa Fadi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
British Telecommunications PLC
Original Assignee
British Telecommunications PLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by British Telecommunications PLC filed Critical British Telecommunications PLC
Priority to GB1718427.6A priority Critical patent/GB2568114B/en
Publication of GB201718427D0 publication Critical patent/GB201718427D0/en
Publication of GB2568114A publication Critical patent/GB2568114A/en
Application granted granted Critical
Publication of GB2568114B publication Critical patent/GB2568114B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/55Detecting local intrusion or implementing counter-measures
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/52Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems during program execution, e.g. stack integrity ; Preventing unwanted data erasure; Buffer overflow
    • G06F21/53Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems during program execution, e.g. stack integrity ; Preventing unwanted data erasure; Buffer overflow by executing in a restricted environment, e.g. sandbox or secure virtual machine
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/45587Isolation or security of virtual machine instances

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Security & Cryptography (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • Computer And Data Communications (AREA)

Abstract

The application defines a system where a machine learning algorithm first receives a plurality of virtual machine configuration vectors for training VMs 602. It then receives a security occurrence identification that identifies a VM configuration vector for a training VM 604, and then defines a vulnerability vector for the identified VM configuration vector 606. This configuration vector along with the vulnerability vector are associated as being a first training example. It then identifies further, earlier VM configuration for the training VM, where each of these further configuration vectors have a temporal indication that precedes that of the referenced VM configuration vector 608. It then uses a reverse decay function to modify the vulnerability vector for each earlier VM configuration 610, and then associates a modified form of the vulnerability vector with each of the further VM configuration vectors as further training examples 612. The vulnerability vectors may include an indicator of each of the security vulnerabilities of the VM.

Description

The present invention relates to the selection of a security policy for a virtualised computer system dynamically.
Virtualised computing environments provide for the instantiation and deployment of potentially multiple virtual machines (VMs) as virtual implementations of physical machines. Improving the security of such VMs is an ongoing challenge as security threats continue to develop taking advantage of potential security vulnerabilities in a VM configuration.
Security information and event management (SIEM) products or services track security events from networks and applications. SIEM can be provided as software solutions or network appliances that log the security events and provide visibility and reporting. These can be combined with a Managed Security Service (MSS) by a MSS Provider (MSSP). An MSSP may manage many VMs which involves managing security policies for those machines.
Security policies for a VM defines a security configuration for the machine and specifies how the security of the machine is managed. A security configuration may include configuration characteristics of the VM such as the availability of certain networking resources, the execution of certain applications, the provision of facilities such as encryption of data at rest and/or in transit, the provision of security facilities such as intrusion detection facilities, firewalls, malware and/or virus detection facilities, remediation facilities and the like, a version, level or patch level of one or more components in the VM, operating system (OS) facilities, a frequency of update/patching and other factors as will be apparent to those skilled in the art.
Security configurations are typically fixed or infrequently adapted and so they can easily become out of date and unreflective of current risks or threats to a VM.
Thus, there is a need to provide improvements to the security configuration of VMs.
The present invention accordingly provides, in a first aspect, a computer implemented method to generate training data for a machine learning algorithm for determining security vulnerabilities of a virtual machine (VM) in a virtualised computing environment, the machine learning algorithm determining the vulnerabilities based on a vector of configuration characteristics for the VM, the method comprising: receiving a plurality of VM configuration vectors for each of one or more training VMs, each configuration vector including attributes of a configuration of a VM and having a temporal indication; receiving a security occurrence identification being referable to a VM configuration vector for a training VM based on a temporal indication of the security occurrence, the security occurrence identification including information for defining a vector of vulnerabilities for the training VM associated with the referenced VM configuration vector, and associating the referenced VM configuration vector with the vulnerability vector as a first training example; identifying one or more further VM configuration vectors for the training VM, each of the further VM configuration vectors having temporal indications preceding that of the referenced VM configuration vector; and associating a modified form of the vulnerability vector with each of the further VM configuration vectors as further training examples, the vulnerability vector being modified for each further VM configuration vector by a reverse decay function such that each temporally earlier VM configuration vector is associated with a vulnerability vector indicating vulnerability to a lesser degree.
Preferably, the reverse decay function is a 0.5kt function based on an increasing time past (t) for a chronologically earlier VM configuration vector and a constant (k).
Preferably, the method further comprises training the machine learning algorithm based on the training examples to classify a configuration vector for a VM to a vulnerability vector.
Preferably, a vector of configuration characteristics includes an indicator of a state of each of a plurality of configuration characteristics for a VM.
Preferably, a vulnerability vector includes an indicator of each of a plurality of security vulnerabilities of a VM.
The present invention accordingly provides, in a second aspect, a computer system including a processor and memory storing computer program code for performing the steps of the methods set out above.
The present invention accordingly provides, in a third aspect, a computer program element comprising computer program code to, when loaded into a computer system and executed thereon, cause the computer to perform the steps of the methods set out above.
Embodiments of the present invention will now be described, by way of example only, with reference to the accompanying drawings, in which:
Figure 1 is a block diagram of a computer system suitable for the operation of embodiments of the present invention;
Figure 2 is a component diagram of an arrangement for selecting a security configuration for a target VM in accordance with embodiments of the present invention;
Figure 3 illustrates an arrangement of the machine learning algorithm of Figure 2 in accordance with embodiments of the present invention;
Figure 4 is a component diagram of an arrangement for training the machine learning algorithm of Figures 2 and 3 in accordance with embodiments of the present invention;
Figure 5 depicts the application of a reverse decay function such that temporally earlier training VM configurations are associated with vulnerability vectors indicating vulnerability to a lesser degree in accordance with embodiments of the present invention;
Figure 6 is a flowchart of a method to generate training data for the machine learning algorithm of Figures 2 and 3 in accordance with embodiments of the present invention;
Figure 7 is a component diagram of an arrangement for determining a security configuration for a target VM in accordance with embodiments of the present invention; and
Figure 8 is a flowchart of a method for determining a security configuration for a target VM in accordance with embodiments of the present invention.
Figure 1 is a block diagram of a computer system suitable for the operation of embodiments of the present invention. A central processor unit (CPU) 102 is communicatively connected to a storage 104 and an input/output (I/O) interface 106 via a data bus 108. The storage 104 can be any read/write storage device such as a random access memory (RAM) or a non-volatile storage device. An example of a non-volatile storage device includes a disk or tape storage device. The I/O interface 106 is an interface to devices for the input or output of data, or for both input and output of data. Examples of I/O devices connectable to I/O interface 106 include a keyboard, a mouse, a display (such as a monitor) and a network connection.
Figure 2 is a component diagram of an arrangement for selecting a security configuration 216 for a target VM in accordance with embodiments of the present invention. The security configuration 216 is one of a set of possible security configurations each aligning to a security policy, and is selected by a security configuration selector 214 as a hardware, software, firmware or combination component adapted to select a security configuration for a target VM based on an evaluated measure of risk. The evaluated measure of risk is determined by a risk evaluator 212 as a hardware, software, firmware or combination component based on parameters including at least: global security threat information 206 such as may be provided by third party entities such as security monitoring agencies, attack detection and monitoring agencies and the like; local security threat information 208 such as may be determined by local monitoring and/or threat determination measures such as security facilities of a virtualised computing environment, network, local appliances or the like; and a vulnerability vector 210 for the target VM, the vulnerability vector being determined by a machine learning method based on a configuration of the target VM as described below. Thus, in use, the risk evaluator 212 evaluates a measure of risk for the target VM to inform a determination by the security configuration selector 214 of an appropriate security configuration 216 for application to the VM,
A security configuration 216, as an embodiment of a security policy, can include, inter alia: the installation, deinstallation, modification or configuration of one or more resources for the VM including, for example, software, patches, services, security facilities and the like; the deployment of one or more specific security protective measures such as intrusion detection, virus detection, malware detection, network scanning, real-time scanning, network monitoring, network configuration , or other facilities as will be apparent to those skilled in the art; patching, updating, replacement or reconfiguration of operating system, kernel, hypervisor, memory, storage, processor or other resources of the target VM; and other security configuration measures.
In addition to the threat information 206, 208, the security configuration selector 214 is responsive to the risk evaluation based on a determination of a degree of vulnerability of the target VM. The degree of vulnerability is reflected in the VM vulnerability vector 210 as a vector of metrics each corresponding to an indication of an extent of a weakness of the target VM to exploitation. For example, the opening of networking ports introduces a vulnerability to exploitation of those open ports by malicious entities. The nature, number, extent and security of such open ports can be reflected as one or more metrics in the VM vulnerability vector 210 for the target VM. The risk evaluated by the risk evaluator 212 thus constitutes a metric combining a vulnerability of the target VM with matched threats (local and/or global).
The vulnerability vector 210 is determined based on a supervised machine learning algorithm 204 for inferring a function from labelled training data. The machine learning algorithm 204 is trained using training data consisting of training examples each having an input vector and desired output vector. The machine learning algorithm 204 thus analyses the training data to produce an inferred function which can be used for mapping new examples, including an input vector for the target VM to infer an output vector for the target VM. The implementation of such supervised machine learning algorithms is beyond the scope of this specification and will be apparent to those skilled in the art.
In embodiments of the present invention, the machine learning algorithm 204 is trained to identify a VM vulnerability vector 210 for a VM based on a vector of configuration characteristics of the VM known as a VM configuration vector 202. The VM configuration vector 202 is a vector of metrics each corresponding to one or more configuration characteristics of a VM. For example, the VM configuration vector 202 can encode a VM configuration state including and/or based on, inter alia: log file data; operating system type, version and/or update; applications, services, features or other components installed in the VM; user information for the VM; geographic location of the VM or hardware used for instantiating, supporting, running or managing the VM; network information pertaining to the VM such as a network topology, network connections and the like; and security features installed in the VM such as intrusion detection, malware detection, antivirus, firewall and the like. Thus, in use, the machine learning algorithm 204 determines a VM vulnerability vector 210 for a VM configuration vector 202, as depicted in Figure 3.
Training the supervised machine learning algorithm 204 is necessary. In particular, the machine learning algorithm 204 is trained to generate a VM vulnerability vector 210 for a VM configuration vector 202 to reflect known or realised security occurrences. A security occurrence includes an attack or other security event that corresponds, at least in part, to the realisation of the effect of a security vulnerability existing in a VM having a particular VM configuration. Thus, training examples are required for the machine learning algorithm for which a security occurrence was realised and including a configuration vector 202 for a training VM and a corresponding vulnerability vector 210 in view of the security occurrence. The identification of such training examples is particularly challenging, not least because a large number of training examples is required to provide effective training for the machine learning algorithm 204.
Figure 4 is a component diagram of an arrangement for training the machine learning algorithm 204 of Figures 2 and 3 in accordance with embodiments of the present invention. Embodiments of the present invention draw on existing repositories of VM configurations as a source of training examples for the machine learning algorithm 204. In particular, it is known good practice to gather VM configuration information in “data lakes” 402 - i.e. data storage repositories for storing data in potentially diverse schema or structural forms such as files or binary large objects (blobs). Such data lakes 402 provide an opportunity to gather VM configuration information fortraining VMs. For example, the data lake of Figure 4 includes multiple VM configuration information for a first training VM, VMi, ranging temporally from a first configuration VM1C1 to configuration VMiCa. Similarly, multiple VM configuration information for other VMs is provided such as VMn having configuration information ranging temporally from first configuration VMnCi to configuration VMnCb, and so on. Notably, the temporal range of configuration information means that each set of configuration information for a VM is associated with some temporal indication such as a time of its applicability or similar. Thus, from such configuration information, multiple VM configuration vectors 202 can be generated.
Figure 4 includes a training data labeller 404 as a software, hardware, firmware or combination component for generating labelled training data for a machine learning trainer 408 to train the machine learning algorithm 204. The labeller 404 thus formulates each training example as an input VM configuration vector 202, such as based on configuration information for a VM from the data lake 402, where the training example is labelled by a desired output vector as a VM vulnerability vector 210. The VM vulnerability vector 210 is generated based on security occurrence information 406 received in respect of VMs for which configuration data is available in the data lake 402.
The security occurrence information 406 relates to a security attack or other security event occurring in respect of a training VM and includes time information for identifying a particular configuration in the data lake 402 of the training VM at the time of the security occurrence. Thus, the security occurrence is referable to a VM configuration vector derived based on a timed VM configuration in the data lake. Further, the security occurrence information 406 serves as a basis for defining a VM vulnerability vector 210 for the training VM because the security occurrence information 406 includes sufficient information to populate a VM vulnerability vector. Thus, by way of the security occurrence information 406, the training data labeller 404 generates labelled training examples each as a VM configuration vector 202 from the data lake 402 identified and labelled by a VM vulnerability vector derived from the security occurrence information 406.
The security occurrence information 406 relates directly to discrete VM configurations in the data lake based on the timing of the security occurrence and the temporal indication for each VM configuration, and therefore, to discrete VM configuration vectors 202. Thus there will be many (probably a vast majority) VM configurations derivable from the data lake 402 for which no security occurrence information is available and, therefore, for which no VM vulnerability vector 210 can be determined. For these VM configurations, the training data labeller 404 employs a reverse decay strategy to apply a security occurrence 406 to temporally previous VM configuration vectors 202 for the same VM. That is to say that VM configurations occurring prior to a configuration at which a security occurrence 406 was observed can be presumed to contribute to the vulnerability of the VM at the time of the security occurrence 406. In this regard, such prior VM configurations can be considered to have a “pre-vulnerability”. The distance, temporally, that such a presumption can be applied is configurable such that only configurations occurring temporally prior but temporally “close” (according to a formal definition) to that for which the security occurrence 406 occurred are so labelled. The presumption is appropriate because VM configuration is an evolutionary process such that configuration changes occurring over time are cumulative. Accordingly, a single configuration known to be demonstrably vulnerable by virtue of a security occurrence 406 is reasonably considered to impart at least a degree of vulnerability on temporally previous configurations, at least within a configurable or predetermined temporal window.
Figure 5 depicts the application of a reverse decay function such that temporally earlier training VM configurations are associated with vulnerability vectors indicating vulnerability to a lesser degree in accordance with embodiments of the present invention. As can be seen in the graph of Figure 5, a security occurrence 406 is known to have occurred in respect of a security configuration VMiCt applicable at a time, T. Thus, vulnerability of relevant features in the VM vulnerability vector 210 can be considered maxima or unity for this particular VM configuration. Accordingly, the graph shows that such a VM configuration has a degree of vulnerability (for a discrete vulnerability in the VM vulnerability vector 210) of 1.0.
Drawing back from this time T temporally, a previous VM configuration occurring at time T-d (some time delta d prior to 7), VMiC-r-d, is labelled as a training example with a modified form of the VM vulnerability vector for VMiCt applicable at a time, T. In particular, as illustrated in the graph of Figure 5, the degree of vulnerability indicated in the modified vulnerability vector for VMiC-r-d is some degree lower than that for VMiCt. Similarly, a further prior VM configuration occurring at T-2d, VMiCT-2d, is labelled as a training example with a further modified form of the VM vulnerability vector for VMiCt. Thus, a modified form of the vulnerability vector for VMiCt is used to label training examples for multiple preceding VM configurations. Preferably, the vulnerability vector is modified for each further VM configuration vector by a reverse decay function such that each temporally earlier VM configuration vector is associated with a vulnerability vector indicating vulnerability to a lesser degree. In a preferred embodiment, the reverse decay function is 0.5w where t is an increasing time past for a chronologically earlier VM configuration vector and k is a constant for governing the vulnerability duration - i.e. how far temporally back from a security occurrence configurations are to be employed as training examples. Note that the increasing value of t for chronologically earlier VM configurations with each earlier VM configuration being labelled by an increasingly decayed degree of vulnerability provides the reverse decay function such that decay increases backwards chronologically for earlier VM configurations.
Thus, in this way, a number of useful training examples for the machine learning algorithm 204 is increased by interpolating training examples from prior VM configurations based on security occurrence information. Notably, the nature of the security occurrence information 406 can itself lead to a decay or reduction of a degree of vulnerability even for a VM configuration applicable at the time of the occurrence, depending on the type of occurrence. For example, where the occurrence 406 is a confirmed security attack, then no reduction in the degree of vulnerability may apply. However, where the occurrence 406 is one of a number of security events of potentially differing severity, such as a virus detection and removal, a suspicion of intrusion and the like, then the degree of vulnerability reflects in the VM vulnerability vector can be reduced according to predetermined measures or factors so as to emphasise vulnerability of a VM configuration in the event of a confirmed attack and to relatively de-emphasise vulnerability in the event of less significant security events.
Figure 6 is a flowchart of a method to generate training data for the machine learning algorithm of Figures 2 and 3 in accordance with embodiments of the present invention. Initially, at step 602, a plurality of VM configuration vectors 202 for each of one or more training VMs are received. Each received configuration vector 202 includes attributes of a configuration of a training VM and has a temporal indication such as a time of applicability of the VM configuration corresponding to the configuration vector 202. At step 604, a security occurrence identification 406 is received, the security occurrence identification 406 being referable to a VM configuration vector 202 for a training VM based on a time or other temporal indication of the security occurrence 406. The security occurrence identification 406 includes information for defining a VM vulnerability vector 210 for the training VM associated with the VM configuration vector 202. At step 606, the referenced VM configuration vector 202 is associated with (labelled by) the VM vulnerability vector 210 as a first training example. At step 608, one or more further VM configuration vectors 202 for the training VM are identified. Each of the further VM configuration vectors 202 has a temporal indication preceding that of the referenced VM configuration vector such that they applied earlier than the referenced VM configuration. At step 610 a modified form of the vulnerability vector 210 is associated with each of the further VM configuration vectors 202 as further training examples, the vulnerability vector 210 being modified for each further VM configuration vector by a reverse decay function. At step 612 each temporally earlier VM configuration vector 210 is associated with a vulnerability vector indicating vulnerability to a lesser degree.
Figure 7 is a component diagram of an arrangement for determining a security configuration for a target VM 700 in accordance with embodiments of the present invention. Many of the features of Figure 7 are identical to those described above with respect to Figure 2 and these will not be repeated here. The machine learning algorithm 704 of Figure 7 differs from that previously described in that it is trained for VM configuration vectors 202 that include an identification of a security configuration 716 applicable for a VM. These training example VM configuration vectors 202 are labelled by VM vulnerability vectors 210 as previously described, though the inclusion of the security configuration information with the
VM configuration vector 202 encodes the security configuration within the trained machine learning algorithm 704. Subsequently, in use in operation, a VM configuration vector 202 for the target VM can be supplemented by one or more prospective security configurations as input to the machine learning algorithm 704 to elicit a VM vulnerability vector 710 reflecting the prospective security configuration. Thus, in this way, it is possible to sample the machine learning algorithm 704 with a VM configuration vector 202 for the target VM supplemented by each of a number of prospective security configurations 716 (i.e. executing the machine learning algorithm 704 multiple times) to elicit multiple VM vulnerability vectors 710, each reflective of a different prospective security configuration.
The approach of Figure 7 is particularly useful for the selection of a security configuration 730 for a VM in the context of global and/or local threats 706, 708. For example, threats 706, 708 can serve to prioritise the vulnerability features in a VM vulnerability vector 710 such that some vulnerability features are more significant than others in view of the threats posed. For example, a network attack threat involving a particular network protocol can serve to emphasise features in the VM vulnerability vector 710 corresponding to networking, the particular network protocol, and/or applications or features using the network protocol. These features can then be prioritised. Accordingly, executing the machine learning algorithm 704 for a VM configuration vector 202 of the target VM 700 multiple times with different proposed security configurations 716 to elicit multiple VM vulnerability vectors 710, the multiple VM vulnerability vectors can be ranked according to the degree of vulnerability of the prioritised features in order that the security configuration 730 offering the lowest vulnerable in the prioritised features can be selected by a security configuration selector 714.
Figure 8 is a flowchart of a method for determining a security configuration 730 for a target VM 700 in accordance with embodiments of the present invention. Initially, at step 802, the machine learning algorithm 704 is trained to determine a vector 710 of security vulnerabilities for the target VM 700 based on a vector 202 of configuration characteristics for the target VM. The machine learning algorithm is trained using training examples each including a configuration for a training VM and an associated vulnerability vector based on an observed security occurrence 406 at the training VM. In particular, each training example includes an identification of one of set of security configurations 716 for the training VM. At step 804 the method selects at least a subset of the set of security configurations 716 (potentially all of the security configurations). At step 806 the method executes the machine learning algorithm 704 for each security configuration 716 in the selected subset. The machine learning algorithm 704 is executed with a VM configuration vector 202 for the target VM and an identification of each security configuration. The machine learning algorithm 704 therefore generates a set of VM vulnerability vectors 710 including a vulnerability vector for each security configuration 716 in the selected subset. At step 808 the method selects a security configuration 730 for the target VM 700 based on the set of VM vulnerability vectors 710.
Insofar as embodiments of the invention described are implementable, at least in part, using a software-controlled programmable processing device, such as a microprocessor, digital signal processor or other processing device, data processing apparatus or system, it will be appreciated that a computer program for configuring a programmable device, apparatus or system to implement the foregoing described methods is envisaged as an aspect of the present invention. The computer program may be embodied as source code or undergo compilation for implementation on a processing device, apparatus or system or may be embodied as object code, for example.
Suitably, the computer program is stored on a carrier medium in machine or device readable form, for example in solid-state memory, magnetic memory such as disk or tape, optically or magneto-optically readable memory such as compact disk or digital versatile disk etc., and the processing device utilises the program or a part thereof to configure it for operation. The computer program may be supplied from a remote source embodied in a communications medium such as an electronic signal, radio frequency carrier wave or optical carrier wave. Such carrier media are also envisaged as aspects of the present invention.
It will be understood by those skilled in the art that, although the present invention has been described in relation to the above described example embodiments, the invention is not limited thereto and that there are many possible variations and modifications which fall within the scope of the invention.
The scope of the present invention includes any novel features or combination of features disclosed herein. The applicant hereby gives notice that new claims may be formulated to such features or combination of features during prosecution of this application or of any such further applications derived therefrom. In particular, with reference to the appended claims, features from dependent claims may be combined with those of the independent claims and features from respective independent claims may be combined in any appropriate manner and not merely in the specific combinations enumerated in the claims.

Claims (7)

1. A computer implemented method to generate training data for a machine learning algorithm for determining security vulnerabilities of a virtual machine (VM) in a virtualised computing environment, the machine learning algorithm determining the vulnerabilities based on a vector of configuration characteristics for the VM, the method comprising:
receiving a plurality of VM configuration vectors for each of one or more training VMs, each configuration vector including attributes of a configuration of a VM and having a temporal indication;
receiving a security occurrence identification being referable to a VM configuration vector for a training VM based on a temporal indication of the security occurrence, the security occurrence identification including information for defining a vector of vulnerabilities for the training VM associated with the referenced VM configuration vector, and associating the referenced VM configuration vector with the vulnerability vector as a first training example;
identifying one or more further VM configuration vectors for the training VM, each of the further VM configuration vectors having temporal indications preceding that of the referenced VM configuration vector; and associating a modified form of the vulnerability vector with each of the further VM configuration vectors as further training examples, the vulnerability vector being modified for each further VM configuration vector by a reverse decay function such that each temporally earlier VM configuration vector is associated with a vulnerability vector indicating vulnerability to a lesser degree.
2. The method of claim 1 wherein the reverse decay function is a 0.5w function based on an increasing time past (t) for a chronologically earlier VM configuration vector and a constant (k).
3. The method of any preceding claim further comprising training the machine learning algorithm based on the training examples to classify a configuration vector for a VM to a vulnerability vector.
4. The method of any preceding claim wherein a vector of configuration characteristics includes an indicator of a state of each of a plurality of configuration characteristics for a VM.
5. The method of any preceding claim wherein a vulnerability vector includes an indicator of each of a plurality of security vulnerabilities of a VM.
6. A computer system including a processor and memory storing computer program
5 code for performing the steps of a method as claimed in any of claims 1 to 5.
7. A computer program element comprising computer program code to, when loaded into a computer system and executed thereon, cause the computer to perform the steps of a method as claimed in any of claims 1 to 5.
GB1718427.6A 2017-11-07 2017-11-07 Training a machine learning algorithm to define vulnerability vectors for a virtual machine configuration vector Expired - Fee Related GB2568114B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
GB1718427.6A GB2568114B (en) 2017-11-07 2017-11-07 Training a machine learning algorithm to define vulnerability vectors for a virtual machine configuration vector

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB1718427.6A GB2568114B (en) 2017-11-07 2017-11-07 Training a machine learning algorithm to define vulnerability vectors for a virtual machine configuration vector

Publications (3)

Publication Number Publication Date
GB201718427D0 GB201718427D0 (en) 2017-12-20
GB2568114A true GB2568114A (en) 2019-05-08
GB2568114B GB2568114B (en) 2020-05-06

Family

ID=60664652

Family Applications (1)

Application Number Title Priority Date Filing Date
GB1718427.6A Expired - Fee Related GB2568114B (en) 2017-11-07 2017-11-07 Training a machine learning algorithm to define vulnerability vectors for a virtual machine configuration vector

Country Status (1)

Country Link
GB (1) GB2568114B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11132444B2 (en) * 2018-04-16 2021-09-28 International Business Machines Corporation Using gradients to detect backdoors in neural networks

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160239330A1 (en) * 2015-02-12 2016-08-18 Alcatel-Lucent Usa Inc. Dynamic Reconfiguration Of Resources In A Virtualized Network
US20170147816A1 (en) * 2015-11-23 2017-05-25 Armor Defense Inc. Extracting Malicious Instructions on a Virtual Machine
US20170279826A1 (en) * 2016-03-22 2017-09-28 Symantec Corporation Protecting dynamic and short-lived virtual machine instances in cloud environments

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160239330A1 (en) * 2015-02-12 2016-08-18 Alcatel-Lucent Usa Inc. Dynamic Reconfiguration Of Resources In A Virtualized Network
US20170147816A1 (en) * 2015-11-23 2017-05-25 Armor Defense Inc. Extracting Malicious Instructions on a Virtual Machine
US20170279826A1 (en) * 2016-03-22 2017-09-28 Symantec Corporation Protecting dynamic and short-lived virtual machine instances in cloud environments

Also Published As

Publication number Publication date
GB2568114B (en) 2020-05-06
GB201718427D0 (en) 2017-12-20

Similar Documents

Publication Publication Date Title
US11763005B2 (en) Dynamic security policy
US11997097B2 (en) Security vulnerability assessment for users of a cloud computing environment
US12175853B2 (en) Adaptive severity functions for alerts
US11151256B2 (en) Detecting security threats by monitoring chains of configuration changes made to basic input/output system (BIOS) or unified extensible firmware interface (UEFI) attributes
US9495180B2 (en) Optimized resource allocation for virtual machines within a malware content detection system
US10936717B1 (en) Monitoring containers running on container host devices for detection of anomalies in current container behavior
US10121000B1 (en) System and method to detect premium attacks on electronic networks and electronic devices
US10853489B2 (en) Data-driven identification of malicious files using machine learning and an ensemble of malware detection procedures
US11991191B2 (en) Detecting a missing security alert using a machine learning model
US8904531B1 (en) Detecting advanced persistent threats
US9536091B2 (en) System and method for detecting time-bomb malware
US10216934B2 (en) Inferential exploit attempt detection
US11775653B2 (en) Security configuration determination
EP4381405B1 (en) Continuous vulnerability assessment system
CA3172788A1 (en) Endpoint security using an action prediction model
US20190318080A1 (en) Stack Pivot Exploit Detection and Mitigation
GB2568114A (en) Dynamic security policy
GB2568115A (en) Security configuration determination
GB2625749A (en) Predicting attack paths
Arshad et al. A Machine Learning Based Intrusion Impact Analysis Scheme for Clouds

Legal Events

Date Code Title Description
PCNP Patent ceased through non-payment of renewal fee

Effective date: 20231107