+

US20190065236A1 - Ensuring the privacy and integrity of a hypervisor - Google Patents

Ensuring the privacy and integrity of a hypervisor Download PDF

Info

Publication number
US20190065236A1
US20190065236A1 US16/177,084 US201816177084A US2019065236A1 US 20190065236 A1 US20190065236 A1 US 20190065236A1 US 201816177084 A US201816177084 A US 201816177084A US 2019065236 A1 US2019065236 A1 US 2019065236A1
Authority
US
United States
Prior art keywords
hypervisor
memory
operating system
host operating
virtual machine
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/177,084
Inventor
Ian Pratt
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hewlett Packard Development Co LP
Original Assignee
Bromium Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/526,755 external-priority patent/US9021476B1/en
Priority claimed from US13/526,769 external-priority patent/US9135046B1/en
Application filed by Bromium Inc filed Critical Bromium Inc
Priority to US16/177,084 priority Critical patent/US20190065236A1/en
Assigned to BROMIUM, INC. reassignment BROMIUM, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PRATT, IAN
Publication of US20190065236A1 publication Critical patent/US20190065236A1/en
Assigned to HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. reassignment HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BROMIUM, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/468Specific access rights for resources, e.g. using capability register
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/52Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems during program execution, e.g. stack integrity ; Preventing unwanted data erasure; Buffer overflow
    • G06F21/53Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems during program execution, e.g. stack integrity ; Preventing unwanted data erasure; Buffer overflow by executing in a restricted environment, e.g. sandbox or secure virtual machine
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/64Protecting data integrity, e.g. using checksums, certificates or signatures
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5011Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
    • G06F9/5016Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals the resource being the memory
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5077Logical partitioning of resources; Management or configuration of virtualized resources
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/45587Isolation or security of virtual machine instances

Definitions

  • Embodiments of the invention relate to ensuring the privacy and integrity of a hypervisor.
  • a hypervisor is a software component that enables multiple operating systems (termed “guest operating systems”) to run on a single device. Each guest operating system is executed within a separate virtual machine.
  • Type 1 hypervisor also known as a native, bare metal hypervisor
  • a Type 1 hypervisor executes directly upon the hardware of the device.
  • a Type 1 hypervisor a guest operating system runs on a level above the hypervisor. Since a Type 1 hypervisor executes directly upon the hardware, a Type 1 hypervisor is the most privileged software component within the system.
  • a Type 2 hypervisor (also known as a hosted hypervisor) executes within a conventional operating system (termed the “host operating system” in this context).
  • a Type 2 hypervisor instantiates guest operating systems using the memory and resources allocated to the hypervisor by the host operating system.
  • guest operating systems run at a lower privilege level than the hypervisor.
  • a Type 2 hypervisor is not the most privileged component within the system. Instead, in a system employing a Type 2 hypervisor, the host operating system is at the same privilege level as the Type 2 hypervisor, and hence, the host operating system can interfere with the Type 2 hypervisor.
  • Type 2 hypervisor If malicious code infects a host operating system that employs a Type 2 hypervisor, then the malicious could potentially compromise the Type 2 hypervisor, since processes executing within the host operating system kernel are running at the same privilege level as the Type 2 hypervisor. It is desirable to prevent a Type 2 hypervisor (and by extension guest virtual machines) from being compromised from any malicious code executing within the host operating system.
  • FIG. 1 is a block diagram illustrating an exemplary system according to an embodiment of the invention
  • FIG. 2 is a flowchart illustrating the functional steps of saving a virtual machine according to an embodiment of the invention.
  • FIG. 3 is a block diagram that illustrates a computer system upon which an embodiment of the invention may be implemented.
  • Embodiments of the invention protect the privacy and integrity of a hypervisor executing within a host operating system. To do so, embodiments enable the hypervisor to be the most privileged software component operating within a system. For example, using an embodiment, the hypervisor is more privileged than the host operating system. In this way, processes executing outside of the hypervisor (such as those processes executing within the host operating system) cannot affect the privacy or integrity of the hypervisor.
  • the host operating system is prevented from accessing a portion of resources belonging to or allocated by a hypervisor to prevent the host operating system from violating the privacy and integrity of those resources.
  • the host operating system may be prevented from accessing resources belonging to or allocated by the hypervisor by transferring execution of the host operating system into a virtual machine container that does not have sufficient privilege to access any portion of the memory pages in which the hypervisor is executing.
  • resources of the hypervisor may be protected in this manner.
  • resources of the hypervisor include memory, scheduling access to a central processing unit (CPU), and access to one or more input/output devices.
  • CPU central processing unit
  • FIG. 1 is a block diagram illustrating an exemplary system 100 according to an embodiment of the invention.
  • System 100 includes a set of software components, namely host operating system 110 (hereafter “host OS 110 ”), hypervisor 130 , and virtual machines 140 and 142 .
  • the software components of system 100 may reside and operate on a wide variety of physical devices.
  • non-limiting, illustrative examples of the types of physical devices upon which system 100 may reside and operate include a personal computer (PC), a tablet PC, a laptop computer, a cell phone, a personal digital assistance (PDA), and the like.
  • PC personal computer
  • PDA personal digital assistance
  • Host OS 110 is intended to broadly represent any type of operating system which is capable of managing a set of resources.
  • Hypervisor 130 is intended to broadly represent any type of hypervisor that executes within host OS 110 .
  • Hypervisors that execute within a host operating system may be referred to as Type 2 hypervisors; although, as shall be evident from the present teachings, hypervisor 130 of an embodiment operates differently than Type 2 hypervisors known in the art.
  • Hypervisor 130 is capable of creating and managing a plurality of different operating systems (conventionally termed “guest operating systems”) using the memory and other resources available to hypervisor 130 .
  • Guest operating systems within system 100 are intended to execute within a virtual machine.
  • FIG. 1 depicts guest operating systems (hereafter simply guest OS) 150 and 152 which execute within virtual machines 140 and 142 respectively.
  • hypervisor 130 may create any number of virtual machines. Furthermore, hypervisor 130 may, at various times, de-allocate a particular virtual machine when it is no longer needed. Therefore, the number of virtual machines, and by extension the number of guest operating systems, within system 100 may vary over time and from implementation to implementation.
  • Resources such as memory, access to the CPU, and access to an I/O device, may be assigned or allocated to hypervisor 130 .
  • Hypervisor 130 may allocate a certain portion of these resources to a virtual machine instantiated by hypervisor 130 .
  • Each guest operating system may comprise one or more components that issue requests to perform operations (such as an I/O operation) to host OS 110 via hypervisor 130 .
  • Such operations may include a read request, a write request, a request to allocate one or more resources, and the like.
  • FIG. 1 depicts an example of a component in guest OS 150 that issues a request to perform operations to host OS 110 as device driver component 154 .
  • Device driver component 154 may, in certain embodiments, correspond to a paravirtual device driver that is designed specifically for a hypervisor environment or to a hardware device driver designed to communicate with a hardware component, which in a hypervisor environment communicates with an emulation of that hardware device.
  • Device driver component 154 may require access to resources, such as access to a persistent storage, a network, or the CPU, and may issue a request for such resources.
  • Host OS 110 will process requests originating from device driver component 154 for access to resources.
  • the requests for access to a resource issued by device driver component 154 and processed by host OS 110 may take many forms.
  • device driver component 154 may issue requests to read a resource (such as a page of memory or a data block stored on disk for example), write to a resource, or allocate a resource.
  • device driver component 154 may issue a request for additional memory resources, either as an explicit request for more memory or as a result of attempting to read or write to a page of memory in its guest-physical address space that is not currently mapped to any actual memory, as doing so will generate a page fault, and as a consequence, will cause hypervisor 130 to attempt to map the page of memory in its guest-physical address space to a page of actual memory.
  • Virtual machine container 120 refers to any container, abstraction, or mechanism which may be used to enforce and manage access to a bounded area of memory.
  • virtual machine container 120 may be implemented as a VMCS container provided by certain Intel® hardware by an embodiment.
  • Virtual machine container 120 is designed or configured to have insufficient privilege to access any portion of the set of memory pages in which hypervisor 130 , or any virtual machine instantiated thereby, reside. Therefore, processes executing within virtual machine container 120 cannot access resources assigned to or allocated by hypervisor 130 , but for certain exceptions discussed below.
  • Embodiments may employ different mechanisms to ensure that virtual machine container 120 lacks sufficient privilege to access resources allocated to or by hypervisor 130 .
  • One way for doing so is by removing reference to the set of memory pages in which hypervisor 130 resides from the Extended Page Tables (EPT) of host OS 110 .
  • Extended Page Tables are tables used by a Memory Management Unit (MMU) to map guest-physical addresses to host-physical addresses.
  • MMU Memory Management Unit
  • the Extended Page Tables of virtual machine container 120 allows host OS 110 to access the memory and other resources host OS 110 is supposed to possess, but not the memory and other resources belonging to or allocated by hypervisor 130 .
  • Hardware components such as a MMU, may establish and enforce constraints on what portions of memory host OS 110 is allowed to access based on data within the Extended Page Tables.
  • Embodiments may take different approaches for establishing host OS 110 within virtual machine container 120 .
  • host OS 110 boots. Thereafter, after booting host OS 110 , host OS 110 is transferred into a virtual machine container 120 .
  • a portion of machine memory that host OS 110 cannot access is reserved. This reserved area of memory may be a VMCS container in an embodiment.
  • hypervisor 130 is instantiated within the reserved portion of memory which host OS 110 cannot access.
  • host OS 110 once host OS 110 is prevented from accessing any resources belonging to or allocated by hypervisor 130 , devices under the control of host OS 110 are prevented from performing unauthorized direct memory accesses (DMA) to resources belonging to or allocated by hypervisor 130 by means of configuring the input/output memory management unit (IOMMU).
  • DMA direct memory accesses
  • IOMMU input/output memory management unit
  • hypervisor 130 may provide host 110 temporary access to those memory pages involved in the performance of an operation (such as an I/O operation). Thereafter, hypervisor 130 may revoke access to those memory pages from host OS 110 , presumably once the I/O operation has been performed. In this way, host OS 110 only has access to those memory pages necessary to perform the operations it has been requested to perform by hypervisor 130 or virtual machines 140 and 142 .
  • hypervisor 130 or virtual machines 140 and 142 may provide host OS 110 temporary access to a set of memory pages by granting a token to host OS 110 .
  • host OS 110 performs an I/O operation it can present the token to the hypervisor 130 to facilitate the reading from or writing to the location of memory identified by the token as required by the requested operation.
  • host OS 110 need only provide a token.
  • host OS 110 wishes to read or write to a set of memory pages allocated to hypervisor 130
  • host OS 110 notifies hypervisor 130 that host OS 110 desires that a particular operation be performed against a set of memory pages and provides a token for those memory pages to hypervisor 130 . If the token is valid, then hypervisor 130 will permit the read or write operation to the set of memory pages identified by the token. Hypervisor 130 will subsequently revoke the temporary permission to write to that location in memory from host OS 110 by causing the token to become invalid. If the token is invalid, then hypervisor 130 will not permit the operation to be performed by host OS 110 .
  • a guest operating system uses a paravirtual device driver
  • the paravirtual device driver will make an explicit request to hypervisor 130 to allow host OS 110 access to a particular area of memory for purposes of performing 10 operations.
  • hypervisor 130 will have to parse the instructions given to device driver component 154 for the emulated hardware to identify the areas of memory in which the guest operating system is implicitly desiring IO operations to be performed against.
  • hypervisor 130 When hypervisor 130 or a guest OS desires an I/O operation to be performed, hypervisor 130 will ask host OS 110 to perform the I/O operation. As explained above, in doing so, hypervisor 130 will provide a token to host OS 110 so that host OS will have temporary access to a set of memory pages in which the I/O operation is to be performed.
  • hypervisor 130 or a guest operating system may encrypt data before the data is provided to host OS 110 so that host OS 110 may perform an I/O operation using an encrypted version of the data rather than the original version. In this way, host OS 110 only has access to an encrypted version of the data and any programs executing within host OS 110 cannot inspect the encrypted version of the data. As a result, the privacy of hypervisor 130 and guest OSes is ensured since other components cannot inspect data written or read by hypervisor.
  • hypervisor 130 may encrypt the page of memory to create an encrypted version prior to requesting host OS 110 to write the page of memory to a persistent storage, such as a hard-disk drive (HDD) or solid state device (SSD).
  • a guest operating system executing within a virtual machine such as guest OS 150 of virtual machine 140 , may request hypervisor 130 to perform an I/O operation, and so hypervisor 130 , in instructing host OS 110 to perform an I/O operation, may be acting on behalf of a virtual machine in system 100 .
  • hypervisor 130 When hypervisor 130 requests host OS 110 to read into memory the page of memory from persistent storage, host OS 110 will read the encrypted page of memory into a location in memory using the token provided to host OS 110 by hypervisor 130 .
  • hypervisor 130 decrypts the encrypted version of the page of memory. If a virtual machine requested the performance of the read operation, then hypervisor 130 would then provide that virtual machine access to the page of memory.
  • hypervisor 130 copies the encrypted version into a different location in memory to which the host OS 110 does not have permission to access.
  • Hypervisor 130 may copy the encrypted version into the different location in memory either of its own initiative or at the direction of a guest operating system, or as a result of the host presenting a suitable token.
  • hypervisor 130 decrypts the encrypted version of the page of memory at its present location. If a virtual machine requested the performance of the read operation, then hypervisor 130 would then provide that virtual machine access to the page of memory.
  • hypervisor 130 may use a cryptographic checksum to ensure data outside the control of hypervisor 130 has not been tampered with. Using a cryptographic checksum, the integrity of resources belonging to or allocated by hypervisor 130 may be ensured.
  • a cryptographic checksum is data that is computed using a set of input data and an algorithm to ensure that the input data has not been modified since the cryptographic checksum was calculated.
  • Embodiments of the invention may employ a wide variety of approaches for obtaining a cryptographic checksum.
  • Various entities may generate the cryptographic checksum, e.g., in some embodiments, hypervisor 130 may generate the cryptographic checksum while in other embodiments a guest operating system or other component may generate the cryptographic checksum and provide it to hypervisor 130 .
  • Embodiments of the invention may obtain a cryptographic checksum upon encrypted or unencrypted data.
  • hypervisor 130 instructing host OS 110 to write a page of memory to persistent storage.
  • hypervisor 130 generates a cryptographic checksum on a page of memory prior to providing host OS 130 access to the encrypted version of the page of memory.
  • the cryptographic checksum will be used by hypervisor 130 to verify the integrity of the encrypted version of the page of memory.
  • Embodiment of the invention may obtain the cryptographic checksum on the page of memory either before or after encrypting the page of memory.
  • Host OS 110 may then write the cryptographic checksum to a persistent storage contemporaneous with writing the encrypted version of the page of memory to the persistent storage.
  • the hypervisor 130 may determine, using a cryptographic checksum, whether the integrity of the encrypted version of a page has been violated since leaving the custody of hypervisor 130 .
  • the cryptographic checksum may be retrieved by host OS 110 from the persistent storage on which the encrypted version of the data was stored.
  • Host OS 110 may present the cryptographic checksum to hypervisor 130 along with the encrypted version of the data.
  • Embodiments of the invention enable hypervisor 130 to save and restore entire virtual machines in a manner that preserves the integrity and privacy of the virtual machine.
  • FIG. 2 is a flowchart illustrating the functional steps of saving a virtual machine according to an embodiment of the invention.
  • hypervisor 130 halts execution of particular virtual machine. For example, hypervisor 130 may halt the execution of virtual machine 140 of FIG. 1 .
  • step 220 hypervisor 130 encrypts the contents of memory associated with virtual machine 140 to create encrypted memory contents. Note that the performance of step 220 is optional, as in certain embodiments the contents of memory associated with virtual machine 140 may be stored on disk in a non-encrypted fashion.
  • step 230 hypervisor 130 generates or obtains one or more cryptographic checksums for either the encrypted contents of memory for virtual machine 140 (if step 220 is performed) or non-encrypted contents of memory for virtual machine 140 (if step 230 is not performed). Note that step 230 may be performed upon either the unencrypted or encrypted memory contents; as a result, in certain embodiments step 230 may be performed before 220 .
  • step 240 host OS 110 stores the memory contents (in an encrypted or non-encrypted fashion) and any cryptographic checksums generated in step 130 on a persistent storage, such as a hard-disk drive (HDD) or a solid state device (SSD).
  • a persistent storage such as a hard-disk drive (HDD) or a solid state device (SSD).
  • a virtual machine Once a virtual machine has been saved to a persistent storage, it may be restored to memory to facilitate the resumption of its operation.
  • host OS 110 reads the memory contents (which may or may not be encrypted) for the virtual machine from the persistent storage into the memory of a newly instantiated virtual machine container.
  • hypervisor 130 may, in any order: (a) measure the integrity of the memory contents using the cryptographic checksum and (b) decrypt, if necessary, the memory contents for the particular virtual machine.
  • host OS 110 may copy the memory contents into a location in memory to which host OS 110 cannot access. Thereafter, the integrity and privacy of the memory contents may be assessed, and absent any corruption or intrusion to the memory contents, the requesting party may safely access the memory contents. Saving and restoring a virtual machine in this fashion may be used by certain embodiments in starting up each virtual machine. In other words, each new virtual machine may be a restore of a previously saved virtual machine. In this way, a newly instantiated virtual machine may have a desired set of characteristics.
  • the desired memory contents of a newly instantiated virtual machine may be comprised within hypervisor 130 ; when a new virtual machine is started, the memory contents for that virtual machine may be supplied by hypervisor 130 .
  • the memory contents copied into memory may correspond to an initial state of memory for a restored virtual machine prior to booting the virtual machine.
  • This initial state of memory may comprises a BIOS or UEFI firmware, and possibly other firmware, and a set of configuration data to ensure the newly instantiated virtual machine may have a desired set of characteristics.
  • Hypervisor 130 may determine that it is necessary to reduce its memory consumption. Alternately, host OS 110 may request hypervisor 130 to reduce the amount of memory used by hypervisor 130 . Hypervisor 130 will evict memory pages used by itself or a virtual machine to reduce the amount of memory consumed by hypervisor 130 . When hypervisor 130 decides to evict a set of memory pages from a particular virtual machine, hypervisor 130 encrypts the set of memory pages and obtains a cryptographic checksum for the encrypted memory pages. Hypervisor 130 then instructs host OS 130 to persistently store the encrypted memory pages and the cryptographic checksum to a persistent storage. Thereafter, hypervisor may safely evict the set of memory pages from memory and reduce, if desired, an amount of memory allocated to hypervisor 130 , and thus return the page to the host.
  • the integrity of hypervisor 130 may be verified loading hypervisor 130 into memory.
  • host OS 110 may relinquish access to the memory belonging to hypervisor 130 .
  • a cryptographic measurement of hypervisor 130 may be obtained using hardware support.
  • Hypervisor 130 may participate in attesting a measured state of a restored virtual machine (regardless of whether the restored virtual machine was stored on disk in an encrypted or non-encrypted fashion) alongside its own state to a third party.
  • the measured state of hypervisor 130 may be recorded in the Trusted Platform Module, which is a dedicated microprocessor designed to secure hardware by integrating cryptographic keys into devices.
  • FIG. 3 is a block diagram that illustrates a computer system 300 upon which an embodiment of the invention may be implemented.
  • computer system 300 includes processor 304 , main memory 306 , ROM 308 , storage device 310 , and communication interface 318 .
  • Computer system 300 includes at least one processor 304 for processing information.
  • Computer system 300 also includes a main memory 306 , such as a random access memory (RAM) or other dynamic storage device, for storing information and instructions to be executed by processor 304 .
  • Main memory 306 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 304 .
  • Computer system 300 further includes a read only memory (ROM) 308 or other static storage device for storing static information and instructions for processor 304 .
  • a storage device 310 such as a magnetic disk or optical disk, is provided for storing information and instructions.
  • Computer system 300 may be coupled to a display 312 , such as a cathode ray tube (CRT), a LCD monitor, and a television set, for displaying information to a user.
  • a display 312 such as a cathode ray tube (CRT), a LCD monitor, and a television set, for displaying information to a user.
  • An input device 314 is coupled to computer system 300 for communicating information and command selections to processor 304 .
  • Other non-limiting, illustrative examples of input device 314 include a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 304 and for controlling cursor movement on display 312 . While only one input device 314 is depicted in FIG. 3 , embodiments of the invention may include any number of input devices 314 coupled to computer system 300 .
  • Embodiments of the invention are related to the use of computer system 300 for implementing the techniques described herein. According to one embodiment of the invention, those techniques are performed by computer system 300 in response to processor 304 executing one or more sequences of one or more instructions contained in main memory 306 . Such instructions may be read into main memory 306 from another machine-readable medium, such as storage device 310 . Execution of the sequences of instructions contained in main memory 306 causes processor 304 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions to implement embodiments of the invention. Thus, embodiments of the invention are not limited to any specific combination of hardware circuitry and software.
  • machine-readable storage medium refers to any tangible medium that participates in storing instructions which may be provided to processor 304 for execution. Such a medium may take many forms, including but not limited to, non-volatile media and volatile media.
  • Non-volatile media includes, for example, optical or magnetic disks, such as storage device 310 .
  • Volatile media includes dynamic memory, such as main memory 306 .
  • Non-limiting, illustrative examples of machine-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, or any other magnetic medium, a CD-ROM, any other optical medium, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, or any other medium from which a computer can read.
  • Various forms of machine readable media may be involved in carrying one or more sequences of one or more instructions to processor 304 for execution.
  • the instructions may initially be carried on a magnetic disk of a remote computer.
  • the remote computer can load the instructions into its dynamic memory and send the instructions over a network link 320 to computer system 300 .
  • Communication interface 318 provides a two-way data communication coupling to a network link 320 that is connected to a local network.
  • communication interface 318 may be an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of telephone line.
  • ISDN integrated services digital network
  • communication interface 318 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN.
  • LAN local area network
  • Wireless links may also be implemented.
  • communication interface 318 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
  • Network link 320 typically provides data communication through one or more networks to other data devices.
  • network link 320 may provide a connection through a local network to a host computer or to data equipment operated by an Internet Service Provider (ISP).
  • ISP Internet Service Provider
  • Computer system 300 can send messages and receive data, including program code, through the network(s), network link 320 and communication interface 318 .
  • a server might transmit a requested code for an application program through the Internet, a local ISP, a local network, subsequently to communication interface 318 .
  • the received code may be executed by processor 304 as it is received, and/or stored in storage device 310 , or other non-volatile storage for later execution.

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Security & Cryptography (AREA)
  • Computer Hardware Design (AREA)
  • Health & Medical Sciences (AREA)
  • Bioethics (AREA)
  • General Health & Medical Sciences (AREA)
  • Storage Device Security (AREA)

Abstract

Approaches for ensuring the privacy and integrity of a hypervisor. A host operating system manages a set of resources. The host operating system is prevented from accessing a portion of the resources belonging to or allocated by the hypervisor. The host operating system may be prevented from accessing resources belonging to or allocated by the hypervisor by transferring execution of the host operating system into a virtual machine container that does not have sufficient privilege to access any portion of the memory pages in which the hypervisor is executing. After the host operating system provides a requested resource to the hypervisor, the hypervisor may use a hardware component that establishes and enforces constraints on what portions of memory the host operating system is allowed to access to protect the requested resource from the host operating system.

Description

    CLAIM OF PRIORITY AND RELATED APPLICATION DATA
  • This application is a continuation of, and claims priority to, U.S. patent application Ser. No. 14/741,147, entitled Ensuring the Privacy and Integrity of a Hypervisor, filed Jun. 16, 2015, invented by Ian Pratt, the disclosure of which is incorporated by reference for all purposes as if fully set forth herein.
  • U.S. patent application Ser. No. 14/741,147 is a continuation-in-part of, and claims priority to, U.S. patent application Ser. No. 13/526,755, U.S. Pat. No. 9,021,476, entitled Ensuring the Privacy and Integrity of a Hypervisor, filed Jun. 19, 2012, invented by Ian Pratt, the disclosure of which is incorporated by reference for all purposes as if fully set forth herein.
  • U.S. patent application Ser. No. 14/741,147 is also a continuation-in-part of, and claims priority to, U.S. patent application Ser. No. 13/526,769, U.S. Pat. No. 9,135,046, entitled Securing a Controller of a Device from a Host Operating System, filed Jun. 19, 2012, invented by Ian Pratt, the disclosure of which is incorporated by reference for all purposes as if fully set forth herein.
  • FIELD OF THE INVENTION
  • Embodiments of the invention relate to ensuring the privacy and integrity of a hypervisor.
  • BACKGROUND
  • A hypervisor is a software component that enables multiple operating systems (termed “guest operating systems”) to run on a single device. Each guest operating system is executed within a separate virtual machine.
  • There are two widely recognized types of hypervisors, termed Type 1 and Type 2. A Type 1 hypervisor (also known as a native, bare metal hypervisor) executes directly upon the hardware of the device. With a Type 1 hypervisor, a guest operating system runs on a level above the hypervisor. Since a Type 1 hypervisor executes directly upon the hardware, a Type 1 hypervisor is the most privileged software component within the system.
  • A Type 2 hypervisor (also known as a hosted hypervisor) executes within a conventional operating system (termed the “host operating system” in this context). A Type 2 hypervisor instantiates guest operating systems using the memory and resources allocated to the hypervisor by the host operating system. As with a Type 1 hypervisor, guest operating systems run at a lower privilege level than the hypervisor. However, unlike a Type 1 hypervisor, a Type 2 hypervisor is not the most privileged component within the system. Instead, in a system employing a Type 2 hypervisor, the host operating system is at the same privilege level as the Type 2 hypervisor, and hence, the host operating system can interfere with the Type 2 hypervisor.
  • If malicious code infects a host operating system that employs a Type 2 hypervisor, then the malicious could potentially compromise the Type 2 hypervisor, since processes executing within the host operating system kernel are running at the same privilege level as the Type 2 hypervisor. It is desirable to prevent a Type 2 hypervisor (and by extension guest virtual machines) from being compromised from any malicious code executing within the host operating system.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Embodiments of the invention are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:
  • FIG. 1 is a block diagram illustrating an exemplary system according to an embodiment of the invention;
  • FIG. 2 is a flowchart illustrating the functional steps of saving a virtual machine according to an embodiment of the invention; and
  • FIG. 3 is a block diagram that illustrates a computer system upon which an embodiment of the invention may be implemented.
  • DETAILED DESCRIPTION OF THE INVENTION
  • Approaches for ensuring the privacy and integrity of a hypervisor are presented herein. In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the embodiments of the invention described herein. It will be apparent, however, that the embodiments of the invention described herein may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form or discussed at a high level in order to avoid unnecessarily obscuring teachings of embodiments of the invention.
  • Functional Overview
  • Embodiments of the invention protect the privacy and integrity of a hypervisor executing within a host operating system. To do so, embodiments enable the hypervisor to be the most privileged software component operating within a system. For example, using an embodiment, the hypervisor is more privileged than the host operating system. In this way, processes executing outside of the hypervisor (such as those processes executing within the host operating system) cannot affect the privacy or integrity of the hypervisor.
  • In an embodiment, to secure the hypervisor, the host operating system is prevented from accessing a portion of resources belonging to or allocated by a hypervisor to prevent the host operating system from violating the privacy and integrity of those resources. In an embodiment, the host operating system may be prevented from accessing resources belonging to or allocated by the hypervisor by transferring execution of the host operating system into a virtual machine container that does not have sufficient privilege to access any portion of the memory pages in which the hypervisor is executing.
  • A wide variety of resources of the hypervisor may be protected in this manner. For example, non-limiting, illustrative examples of the types of resources of the hypervisor which may be protected in this manner include memory, scheduling access to a central processing unit (CPU), and access to one or more input/output devices.
  • Preventing the Host from Accessing Hypervisor Resources
  • FIG. 1 is a block diagram illustrating an exemplary system 100 according to an embodiment of the invention. System 100 includes a set of software components, namely host operating system 110 (hereafter “host OS 110”), hypervisor 130, and virtual machines 140 and 142. The software components of system 100 may reside and operate on a wide variety of physical devices. For example, non-limiting, illustrative examples of the types of physical devices upon which system 100 may reside and operate include a personal computer (PC), a tablet PC, a laptop computer, a cell phone, a personal digital assistance (PDA), and the like.
  • Host OS 110 is intended to broadly represent any type of operating system which is capable of managing a set of resources. Hypervisor 130 is intended to broadly represent any type of hypervisor that executes within host OS 110. Hypervisors that execute within a host operating system may be referred to as Type 2 hypervisors; although, as shall be evident from the present teachings, hypervisor 130 of an embodiment operates differently than Type 2 hypervisors known in the art.
  • Hypervisor 130 is capable of creating and managing a plurality of different operating systems (conventionally termed “guest operating systems”) using the memory and other resources available to hypervisor 130. Guest operating systems within system 100 are intended to execute within a virtual machine. FIG. 1 depicts guest operating systems (hereafter simply guest OS) 150 and 152 which execute within virtual machines 140 and 142 respectively.
  • While FIG. 1 only depicts two virtual machines, it will be evident to those in the art that hypervisor 130 may create any number of virtual machines. Furthermore, hypervisor 130 may, at various times, de-allocate a particular virtual machine when it is no longer needed. Therefore, the number of virtual machines, and by extension the number of guest operating systems, within system 100 may vary over time and from implementation to implementation.
  • Resources, such as memory, access to the CPU, and access to an I/O device, may be assigned or allocated to hypervisor 130. Hypervisor 130, in turn, may allocate a certain portion of these resources to a virtual machine instantiated by hypervisor 130.
  • Each guest operating system may comprise one or more components that issue requests to perform operations (such as an I/O operation) to host OS 110 via hypervisor 130. Such operations may include a read request, a write request, a request to allocate one or more resources, and the like. FIG. 1 depicts an example of a component in guest OS 150 that issues a request to perform operations to host OS 110 as device driver component 154. Device driver component 154 may, in certain embodiments, correspond to a paravirtual device driver that is designed specifically for a hypervisor environment or to a hardware device driver designed to communicate with a hardware component, which in a hypervisor environment communicates with an emulation of that hardware device.
  • Device driver component 154 may require access to resources, such as access to a persistent storage, a network, or the CPU, and may issue a request for such resources. Host OS 110 will process requests originating from device driver component 154 for access to resources.
  • The requests for access to a resource issued by device driver component 154 and processed by host OS 110 may take many forms. For example, device driver component 154 may issue requests to read a resource (such as a page of memory or a data block stored on disk for example), write to a resource, or allocate a resource. As another example, device driver component 154 may issue a request for additional memory resources, either as an explicit request for more memory or as a result of attempting to read or write to a page of memory in its guest-physical address space that is not currently mapped to any actual memory, as doing so will generate a page fault, and as a consequence, will cause hypervisor 130 to attempt to map the page of memory in its guest-physical address space to a page of actual memory.
  • In an embodiment, to prevent host OS 110 from accessing resources belonging to or allocated by hypervisor 130, execution of host OS 110 is transferred into virtual machine container 120. Virtual machine container 120, as broadly used herein, refers to any container, abstraction, or mechanism which may be used to enforce and manage access to a bounded area of memory. For example, virtual machine container 120 may be implemented as a VMCS container provided by certain Intel® hardware by an embodiment. Virtual machine container 120 is designed or configured to have insufficient privilege to access any portion of the set of memory pages in which hypervisor 130, or any virtual machine instantiated thereby, reside. Therefore, processes executing within virtual machine container 120 cannot access resources assigned to or allocated by hypervisor 130, but for certain exceptions discussed below.
  • Embodiments may employ different mechanisms to ensure that virtual machine container 120 lacks sufficient privilege to access resources allocated to or by hypervisor 130. One way for doing so is by removing reference to the set of memory pages in which hypervisor 130 resides from the Extended Page Tables (EPT) of host OS 110. Extended Page Tables are tables used by a Memory Management Unit (MMU) to map guest-physical addresses to host-physical addresses. Thus, in an embodiment, the Extended Page Tables of virtual machine container 120 allows host OS 110 to access the memory and other resources host OS 110 is supposed to possess, but not the memory and other resources belonging to or allocated by hypervisor 130. Hardware components, such as a MMU, may establish and enforce constraints on what portions of memory host OS 110 is allowed to access based on data within the Extended Page Tables.
  • Embodiments may take different approaches for establishing host OS 110 within virtual machine container 120. For example, in one embodiment, initially, host OS 110 boots. Thereafter, after booting host OS 110, host OS 110 is transferred into a virtual machine container 120. In a different approach, prior to booting host OS 110, a portion of machine memory that host OS 110 cannot access is reserved. This reserved area of memory may be a VMCS container in an embodiment. Thereafter, hypervisor 130 is instantiated within the reserved portion of memory which host OS 110 cannot access.
  • In an embodiment, once host OS 110 is prevented from accessing any resources belonging to or allocated by hypervisor 130, devices under the control of host OS 110 are prevented from performing unauthorized direct memory accesses (DMA) to resources belonging to or allocated by hypervisor 130 by means of configuring the input/output memory management unit (IOMMU). However, that is not to say that host OS 110, or devices under control of host OS 110, can never access the resources belonging to or allocated by hypervisor 130. Indeed, in the natural course of operation, hypervisor 130 will rely upon host OS 110 to perform certain activity, such as I/O operations, on behalf of hypervisor 130 and virtual machines 140 and 142.
  • To balance the need to have host OS 110 perform I/O operations for hypervisor 130 while minimizing the scope of access afforded to host OS 110, in embodiments of the invention, hypervisor 130 may provide host 110 temporary access to those memory pages involved in the performance of an operation (such as an I/O operation). Thereafter, hypervisor 130 may revoke access to those memory pages from host OS 110, presumably once the I/O operation has been performed. In this way, host OS 110 only has access to those memory pages necessary to perform the operations it has been requested to perform by hypervisor 130 or virtual machines 140 and 142.
  • To illustrate a concrete example, hypervisor 130 or virtual machines 140 and 142 may provide host OS 110 temporary access to a set of memory pages by granting a token to host OS 110. Thus, when host OS 110 performs an I/O operation it can present the token to the hypervisor 130 to facilitate the reading from or writing to the location of memory identified by the token as required by the requested operation.
  • Rather than providing a memory address to hypervisor 130 to identify a set of memory pages at which a requested operation is to be performed, host OS 110 need only provide a token. When host OS 110 wishes to read or write to a set of memory pages allocated to hypervisor 130, host OS 110 notifies hypervisor 130 that host OS 110 desires that a particular operation be performed against a set of memory pages and provides a token for those memory pages to hypervisor 130. If the token is valid, then hypervisor 130 will permit the read or write operation to the set of memory pages identified by the token. Hypervisor 130 will subsequently revoke the temporary permission to write to that location in memory from host OS 110 by causing the token to become invalid. If the token is invalid, then hypervisor 130 will not permit the operation to be performed by host OS 110.
  • If a guest operating system uses a paravirtual device driver, then the paravirtual device driver will make an explicit request to hypervisor 130 to allow host OS 110 access to a particular area of memory for purposes of performing 10 operations. If the device driver component 154 is for a regular hardware device, then hypervisor 130 will have to parse the instructions given to device driver component 154 for the emulated hardware to identify the areas of memory in which the guest operating system is implicitly desiring IO operations to be performed against.
  • Ensuring the Privacy and Integrity of Hypervisor Memory
  • When hypervisor 130 or a guest OS desires an I/O operation to be performed, hypervisor 130 will ask host OS 110 to perform the I/O operation. As explained above, in doing so, hypervisor 130 will provide a token to host OS 110 so that host OS will have temporary access to a set of memory pages in which the I/O operation is to be performed. In an embodiment, either hypervisor 130 or a guest operating system may encrypt data before the data is provided to host OS 110 so that host OS 110 may perform an I/O operation using an encrypted version of the data rather than the original version. In this way, host OS 110 only has access to an encrypted version of the data and any programs executing within host OS 110 cannot inspect the encrypted version of the data. As a result, the privacy of hypervisor 130 and guest OSes is ensured since other components cannot inspect data written or read by hypervisor.
  • To illustrate using an embodiment where hypervisor 130 is responsible for encrypting the data, if hypervisor 130 instructs host OS 110 to write a page of memory to disk, then hypervisor 130 may encrypt the page of memory to create an encrypted version prior to requesting host OS 110 to write the page of memory to a persistent storage, such as a hard-disk drive (HDD) or solid state device (SSD). Note that a guest operating system executing within a virtual machine, such as guest OS 150 of virtual machine 140, may request hypervisor 130 to perform an I/O operation, and so hypervisor 130, in instructing host OS 110 to perform an I/O operation, may be acting on behalf of a virtual machine in system 100.
  • When hypervisor 130 requests host OS 110 to read into memory the page of memory from persistent storage, host OS 110 will read the encrypted page of memory into a location in memory using the token provided to host OS 110 by hypervisor 130.
  • In an embodiment, after hypervisor 130 revokes from host OS 110 access to the encrypted version of the page in memory, hypervisor 130 decrypts the encrypted version of the page of memory. If a virtual machine requested the performance of the read operation, then hypervisor 130 would then provide that virtual machine access to the page of memory.
  • In an alternate embodiment, after host OS 110 reads the encrypted page of memory into a set of memory pages, hypervisor 130 copies the encrypted version into a different location in memory to which the host OS 110 does not have permission to access. Hypervisor 130 may copy the encrypted version into the different location in memory either of its own initiative or at the direction of a guest operating system, or as a result of the host presenting a suitable token. Thereafter, hypervisor 130 decrypts the encrypted version of the page of memory at its present location. If a virtual machine requested the performance of the read operation, then hypervisor 130 would then provide that virtual machine access to the page of memory.
  • In an embodiment, hypervisor 130 may use a cryptographic checksum to ensure data outside the control of hypervisor 130 has not been tampered with. Using a cryptographic checksum, the integrity of resources belonging to or allocated by hypervisor 130 may be ensured. A cryptographic checksum is data that is computed using a set of input data and an algorithm to ensure that the input data has not been modified since the cryptographic checksum was calculated. Embodiments of the invention may employ a wide variety of approaches for obtaining a cryptographic checksum. Various entities may generate the cryptographic checksum, e.g., in some embodiments, hypervisor 130 may generate the cryptographic checksum while in other embodiments a guest operating system or other component may generate the cryptographic checksum and provide it to hypervisor 130. Embodiments of the invention may obtain a cryptographic checksum upon encrypted or unencrypted data.
  • To illustrate how a cryptographic checksum is used by an embodiment, consider an example involving hypervisor 130 instructing host OS 110 to write a page of memory to persistent storage. In an embodiment, hypervisor 130 generates a cryptographic checksum on a page of memory prior to providing host OS 130 access to the encrypted version of the page of memory. The cryptographic checksum will be used by hypervisor 130 to verify the integrity of the encrypted version of the page of memory. Embodiment of the invention may obtain the cryptographic checksum on the page of memory either before or after encrypting the page of memory. Host OS 110 may then write the cryptographic checksum to a persistent storage contemporaneous with writing the encrypted version of the page of memory to the persistent storage.
  • Advantageously, the hypervisor 130 may determine, using a cryptographic checksum, whether the integrity of the encrypted version of a page has been violated since leaving the custody of hypervisor 130. The cryptographic checksum may be retrieved by host OS 110 from the persistent storage on which the encrypted version of the data was stored. Host OS 110 may present the cryptographic checksum to hypervisor 130 along with the encrypted version of the data.
  • Saving and Restoring a Virtual Machine
  • Embodiments of the invention enable hypervisor 130 to save and restore entire virtual machines in a manner that preserves the integrity and privacy of the virtual machine. FIG. 2 is a flowchart illustrating the functional steps of saving a virtual machine according to an embodiment of the invention. In step 210, hypervisor 130 halts execution of particular virtual machine. For example, hypervisor 130 may halt the execution of virtual machine 140 of FIG. 1.
  • In step 220, hypervisor 130 encrypts the contents of memory associated with virtual machine 140 to create encrypted memory contents. Note that the performance of step 220 is optional, as in certain embodiments the contents of memory associated with virtual machine 140 may be stored on disk in a non-encrypted fashion.
  • In step 230, hypervisor 130 generates or obtains one or more cryptographic checksums for either the encrypted contents of memory for virtual machine 140 (if step 220 is performed) or non-encrypted contents of memory for virtual machine 140 (if step 230 is not performed). Note that step 230 may be performed upon either the unencrypted or encrypted memory contents; as a result, in certain embodiments step 230 may be performed before 220.
  • In step 240, host OS 110 stores the memory contents (in an encrypted or non-encrypted fashion) and any cryptographic checksums generated in step 130 on a persistent storage, such as a hard-disk drive (HDD) or a solid state device (SSD).
  • Once a virtual machine has been saved to a persistent storage, it may be restored to memory to facilitate the resumption of its operation. In an embodiment, to restore a virtual machine stored on a persistent storage, host OS 110 reads the memory contents (which may or may not be encrypted) for the virtual machine from the persistent storage into the memory of a newly instantiated virtual machine container. After removing access to the memory contents from host OS 110, hypervisor 130 may, in any order: (a) measure the integrity of the memory contents using the cryptographic checksum and (b) decrypt, if necessary, the memory contents for the particular virtual machine.
  • Alternately, in another embodiment, to restore a virtual machine stored on a persistent storage, rather than expressly revoking access from host OS 110 to the location in memory to which the memory contents are copied, host OS 110 may copy the memory contents into a location in memory to which host OS 110 cannot access. Thereafter, the integrity and privacy of the memory contents may be assessed, and absent any corruption or intrusion to the memory contents, the requesting party may safely access the memory contents. Saving and restoring a virtual machine in this fashion may be used by certain embodiments in starting up each virtual machine. In other words, each new virtual machine may be a restore of a previously saved virtual machine. In this way, a newly instantiated virtual machine may have a desired set of characteristics. In another approach, the desired memory contents of a newly instantiated virtual machine may be comprised within hypervisor 130; when a new virtual machine is started, the memory contents for that virtual machine may be supplied by hypervisor 130.
  • In an embodiment, the memory contents copied into memory may correspond to an initial state of memory for a restored virtual machine prior to booting the virtual machine. This initial state of memory may comprises a BIOS or UEFI firmware, and possibly other firmware, and a set of configuration data to ensure the newly instantiated virtual machine may have a desired set of characteristics.
  • Reducing Memory
  • Hypervisor 130 may determine that it is necessary to reduce its memory consumption. Alternately, host OS 110 may request hypervisor 130 to reduce the amount of memory used by hypervisor 130. Hypervisor 130 will evict memory pages used by itself or a virtual machine to reduce the amount of memory consumed by hypervisor 130. When hypervisor 130 decides to evict a set of memory pages from a particular virtual machine, hypervisor 130 encrypts the set of memory pages and obtains a cryptographic checksum for the encrypted memory pages. Hypervisor 130 then instructs host OS 130 to persistently store the encrypted memory pages and the cryptographic checksum to a persistent storage. Thereafter, hypervisor may safely evict the set of memory pages from memory and reduce, if desired, an amount of memory allocated to hypervisor 130, and thus return the page to the host.
  • When hypervisor 130 decides to read a set of evicted memory pages back into the memory of a particular virtual machine, hypervisor 130 instructs host OS 130 to read the memory pages (which will be encrypted) and a cryptographic checksum from a persistent storage. Hypervisor 130 verifies the integrity of the encrypted memory pages using the cryptographic checksum, and decrypts the encrypted memory pages. Hypervisor 130 may then provide access to the decrypted memory pages to the particular virtual machine.
  • Verifying Integrity of the Hypervisor
  • In an embodiment, the integrity of hypervisor 130 may be verified loading hypervisor 130 into memory. In an embodiment, after loading hypervisor 130 into memory, host OS 110 may relinquish access to the memory belonging to hypervisor 130. Thereafter, a cryptographic measurement of hypervisor 130 may be obtained using hardware support.
  • Hypervisor 130 may participate in attesting a measured state of a restored virtual machine (regardless of whether the restored virtual machine was stored on disk in an encrypted or non-encrypted fashion) alongside its own state to a third party. The measured state of hypervisor 130 may be recorded in the Trusted Platform Module, which is a dedicated microprocessor designed to secure hardware by integrating cryptographic keys into devices.
  • Hardware Mechanisms
  • In an embodiment, system 100 of FIG. 1 may be implemented on a computer system. FIG. 3 is a block diagram that illustrates a computer system 300 upon which an embodiment of the invention may be implemented. In an embodiment, computer system 300 includes processor 304, main memory 306, ROM 308, storage device 310, and communication interface 318. Computer system 300 includes at least one processor 304 for processing information. Computer system 300 also includes a main memory 306, such as a random access memory (RAM) or other dynamic storage device, for storing information and instructions to be executed by processor 304. Main memory 306 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 304. Computer system 300 further includes a read only memory (ROM) 308 or other static storage device for storing static information and instructions for processor 304. A storage device 310, such as a magnetic disk or optical disk, is provided for storing information and instructions.
  • Computer system 300 may be coupled to a display 312, such as a cathode ray tube (CRT), a LCD monitor, and a television set, for displaying information to a user. An input device 314, including alphanumeric and other keys, is coupled to computer system 300 for communicating information and command selections to processor 304. Other non-limiting, illustrative examples of input device 314 include a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 304 and for controlling cursor movement on display 312. While only one input device 314 is depicted in FIG. 3, embodiments of the invention may include any number of input devices 314 coupled to computer system 300.
  • Embodiments of the invention are related to the use of computer system 300 for implementing the techniques described herein. According to one embodiment of the invention, those techniques are performed by computer system 300 in response to processor 304 executing one or more sequences of one or more instructions contained in main memory 306. Such instructions may be read into main memory 306 from another machine-readable medium, such as storage device 310. Execution of the sequences of instructions contained in main memory 306 causes processor 304 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions to implement embodiments of the invention. Thus, embodiments of the invention are not limited to any specific combination of hardware circuitry and software.
  • The term “machine-readable storage medium” as used herein refers to any tangible medium that participates in storing instructions which may be provided to processor 304 for execution. Such a medium may take many forms, including but not limited to, non-volatile media and volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage device 310. Volatile media includes dynamic memory, such as main memory 306.
  • Non-limiting, illustrative examples of machine-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, or any other magnetic medium, a CD-ROM, any other optical medium, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, or any other medium from which a computer can read.
  • Various forms of machine readable media may be involved in carrying one or more sequences of one or more instructions to processor 304 for execution. For example, the instructions may initially be carried on a magnetic disk of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a network link 320 to computer system 300.
  • Communication interface 318 provides a two-way data communication coupling to a network link 320 that is connected to a local network. For example, communication interface 318 may be an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 318 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links may also be implemented. In any such implementation, communication interface 318 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
  • Network link 320 typically provides data communication through one or more networks to other data devices. For example, network link 320 may provide a connection through a local network to a host computer or to data equipment operated by an Internet Service Provider (ISP).
  • Computer system 300 can send messages and receive data, including program code, through the network(s), network link 320 and communication interface 318. For example, a server might transmit a requested code for an application program through the Internet, a local ISP, a local network, subsequently to communication interface 318. The received code may be executed by processor 304 as it is received, and/or stored in storage device 310, or other non-volatile storage for later execution.
  • In the foregoing specification, embodiments of the invention have been described with reference to numerous specific details that may vary from implementation to implementation. Thus, the sole and exclusive indicator of what is the invention, and is intended by the applicants to be the invention, is the set of claims that issue from this application, in the specific form in which such claims issue, including any subsequent correction. Any definitions expressly set forth herein for terms contained in such claims shall govern the meaning of such terms as used in the claims. Hence, no limitation, element, property, feature, advantage or attribute that is not expressly recited in a claim should limit the scope of such claim in any way. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims (25)

What is claimed is:
1. One or more non-transitory computer-readable storage mediums storing one or more sequences of instructions, which when executed by one or more processors, cause:
a host operating system managing a set of resources;
preventing the host operating system from accessing a portion of said set of resources comprising a set of memory pages belonging to a hypervisor or allocated by the hypervisor to prevent the host operating system from violating the privacy and integrity of said portion of said set of resources, wherein preventing comprises:
in response to receiving a request, from a virtual machine, to read, write, or allocate one or more memory pages, the hypervisor requesting the one or more memory pages from the host operating system,
the host operating system providing said one or more memory pages managed by the host operating system to the hypervisor, and
after the hypervisor receives said one or more memory pages from the host operating system, the hypervisor using a hardware component that establishes and enforces constraints on what areas of memory the host operating system is allowed to access to protect the one or more memory pages from the host operating system by preventing said host operating system access to an unencrypted version of said one or more memory pages.
2. The one or more non-transitory computer-readable storage mediums of claim 1, wherein preventing the host operating system from accessing said portion comprises:
executing the hypervisor in said set of memory pages; and
establish execution of the host operating system into a virtual machine container that does not have sufficient privilege to access any portion of said set of memory pages.
3. The one or more non-transitory computer-readable storage mediums of claim 1, wherein said set of resources includes one or more of memory, scheduling access to a central processing unit (CPU), and access to one or more input/output devices.
4. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
preventing devices under the control of the host operating system from performing an unauthorized direct memory access (DMA) to said portion of resources.
5. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
after the hypervisor grants the host operating system temporary access to a particular set of one or more memory pages to allow the host operating system to perform I/O operations, the hypervisor revoking access to said particular set of one or more memory pages from the host operating system.
6. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
in response to the hypervisor receiving a request from the virtual machine to write a page of memory to a persistent storage, a guest operating system, executing within the virtual machine, encrypting said page of memory to create an encrypted version prior to granting the host operating system access to the encrypted version; and
the host operating system writing the encrypted version to a persistent storage.
7. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
in response to the hypervisor receiving a request from the virtual machine to write a page of memory to a persistent storage, the hypervisor encrypting said page of memory to create an encrypted version prior to granting the host operating system access to the encrypted version; and
the host operating system writing the encrypted version to a persistent storage.
8. The one or more non-transitory computer-readable storage mediums of claim 7, wherein execution of the one or more sequences of instructions further comprises:
the hypervisor generating a cryptographic checksum on said page of memory prior to providing the host operating system access to said encrypted version; and
the host operating system writing the cryptographic checksum to a persistent storage contemporaneous with the host operating system writing the encrypted version to the persistent storage,
wherein said cryptographic checksum is used by the hypervisor to verify the integrity of said page.
9. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
in response to the hypervisor receiving a request from the virtual machine to read a page of memory from a persistent storage, a host operating system reading an encrypted version of the page of memory from the persistent storage into a location in memory; and
after the hypervisor revokes from the host operating system access to the location in memory, the hypervisor decrypting the encrypted version of the page of memory to the page of memory and providing the virtual machine access to the page of memory.
10. The one or more non-transitory computer-readable storage mediums of claim 9, wherein execution of the one or more sequences of instructions further comprises:
the hypervisor determining whether the integrity of said encrypted version has been violated using a cryptographic checksum that was retrieved by the host operating system from the persistent storage.
11. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
in response to the hypervisor receiving a request from the virtual machine to read a page of memory from a persistent storage, a host operating system reading an encrypted version of the page of memory from the persistent storage into memory;
the hypervisor copying the encrypted version into a location in memory to which the host operating system cannot access; and
after the hypervisor decrypts the encrypted version at the location in memory, the hypervisor providing the virtual machine access to the page of memory.
12. The one or more non-transitory computer-readable storage mediums of claim 11, wherein execution of the one or more sequences of instructions further comprises:
the hypervisor determining whether the integrity of said encrypted version has been violated using a cryptographic checksum that was retrieved by the host operating system from the persistent storage.
13. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
restoring to memory a particular virtual machine stored on a persistent storage by performing:
the host operating system reading said memory contents for the particular virtual machine from the persistent storage medium into memory of a newly instantiated virtual machine container; and
the hypervisor measuring said memory contents using a cryptographic checksum.
14. The one or more non-transitory computer-readable storage mediums of claim 13, wherein execution of the one or more sequences of instructions further comprises:
the hypervisor attesting, to another party, (a) a first measured state of the restored particular virtual machine and (b) a second measured state of the hypervisor and its configuration.
15. The one or more non-transitory computer-readable storage mediums of claim 13, wherein the particular virtual machine, after being restored to memory, corresponds to an initial state of memory for the particular virtual machine prior to booting the particular virtual machine, wherein said initial state of memory for the particular virtual machine comprises a BIOS or UEFI firmware and a set of configuration data.
16. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
after the hypervisor halts execution of a particular virtual machine, the hypervisor (a) encrypting contents of memory associated with the particular virtual machine to create encrypted memory contents; and
the host operating system storing the encrypted memory contents on a persistent storage medium.
17. The one or more non-transitory computer-readable storage mediums of claim 16, wherein execution of the one or more sequences of instructions further comprises:
generating a cryptographic checksum for the contents of memory; and
the host operating system storing the cryptographic checksum on the persistent storage medium.
18. The one or more non-transitory computer-readable storage mediums of claim 17, wherein execution of the one or more sequences of instructions further comprises:
restoring to memory the particular virtual machine stored on a persistent storage by performing:
the host operating system reading the encrypted contents for the particular virtual machine from the persistent storage medium into memory of a newly instantiated virtual machine container;
after removing access to the encrypted contents from the host operating system, the hypervisor (a) decrypting the encrypted contents for the particular virtual machine and (b) assessing the integrity of the encrypted contents using the cryptographic checksum.
19. The one or more non-transitory computer-readable storage mediums of claim 13, wherein execution of the one or more sequences of instructions further comprises:
the hypervisor attesting, to another party, (a) a first measured state of the restored particular virtual machine and (b) a second measured state of the hypervisor.
20. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
upon the hypervisor deciding to evict a set of memory pages from a particular virtual machine, the hypervisor: (a) encrypting the set of memory pages to form encrypted memory pages; and (b) generating a cryptographic checksum for said memory pages; and
the hypervisor instructing the host operating system to persistently store the encrypted memory pages and the cryptographic checksum to a persistent storage.
21. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
upon the hypervisor deciding to read a set of evicted memory pages into memory, the hypervisor (a) instructing the host operating system to read encrypted memory pages and a cryptographic checksum from a persistent storage (b) verifying the integrity of the memory pages using the cryptographic checksum, and (c) decrypting the encrypted memory pages; and
the hypervisor providing access to the decrypted memory pages to the particular virtual machine.
22. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
verifying the integrity of the hypervisor when loading the hypervisor into memory by performing:
the host operating system loading the hypervisor into memory;
the host operating system relinquishing access to the memory belonging to the hypervisor; and
using hardware support for performing a cryptographic measurement of the hypervisor.
23. The one or more non-transitory computer-readable storage mediums of claim 1, wherein execution of the one or more sequences of instructions further comprises:
prior to booting the host operating system, reserving a portion of machine memory which the host operating system cannot access; and
after instantiating a main hypervisor component, the main hypervisor component executing within said portion of machine memory.
24. An apparatus for ensuring the privacy and integrity of a hypervisor from a host operating system, comprising:
one or more processors; and
one or more non-transitory computer-readable mediums storing one or more sequences of instructions, which when executed by the one or more processors, cause:
the host operating system managing a set of resources; and
preventing the host operating system from accessing a portion of said set of resources comprising a set of memory pages belonging to a hypervisor or allocated by the hypervisor to prevent the host operating system from violating the privacy and integrity of said portion of said set of resources,
wherein preventing comprises:
in response to receiving a request, from a virtual machine, to read, write, or allocate one or more memory pages, the hypervisor requesting the one or more memory pages from host operating system,
the host operating system providing said one or more memory pages managed by the host operating system to the hypervisor, and
after the hypervisor receives said one or more memory pages form the host operating system, the hypervisor using a hardware component that establishes and enforces constraints on what areas of memory the host operating system is allowed to access to protect the one or more memory pages from the host operating system by preventing said host operating system access to an unencrypted version of said one or more memory pages.
25. A method for ensuring the privacy and integrity of a hypervisor, comprising:
a host operating system managing a set of resources; and
preventing the host operating system from accessing a portion of said set of resources belonging to or allocated by the hypervisor to prevent the host operating system from violating the privacy and integrity of said portion,
wherein preventing comprises:
in response to receiving a request, from a virtual machine, to read, write, or allocate one or more memory pages, the hypervisor requesting the one or more memory pages from the host operating system,
the host operating system providing said one or more memory pages managed by the host operating system to the hypervisor, and
after the hypervisor receives said one or more memory pages from the host operating system, the hypervisor using a hardware component that establishes and enforces constraints on what areas of memory the host operating system is allowed to access to protect the one or more memory pages from the host operating system by preventing said host operating system access to an unencrypted version of said one or more memory pages.
US16/177,084 2012-06-19 2018-10-31 Ensuring the privacy and integrity of a hypervisor Abandoned US20190065236A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/177,084 US20190065236A1 (en) 2012-06-19 2018-10-31 Ensuring the privacy and integrity of a hypervisor

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US13/526,755 US9021476B1 (en) 2012-06-19 2012-06-19 Ensuring the privacy and integrity of a hypervisor
US13/526,769 US9135046B1 (en) 2012-06-19 2012-06-19 Preventing host operating system from inspecting or modifying data received by hardware controller by moving host operating system into a virtual machine after boot up
US14/741,147 US10140139B1 (en) 2012-06-19 2015-06-16 Ensuring the privacy and integrity of a hypervisor
US16/177,084 US20190065236A1 (en) 2012-06-19 2018-10-31 Ensuring the privacy and integrity of a hypervisor

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/741,147 Continuation US10140139B1 (en) 2012-06-19 2015-06-16 Ensuring the privacy and integrity of a hypervisor

Publications (1)

Publication Number Publication Date
US20190065236A1 true US20190065236A1 (en) 2019-02-28

Family

ID=64315444

Family Applications (2)

Application Number Title Priority Date Filing Date
US14/741,147 Active 2033-05-15 US10140139B1 (en) 2012-06-19 2015-06-16 Ensuring the privacy and integrity of a hypervisor
US16/177,084 Abandoned US20190065236A1 (en) 2012-06-19 2018-10-31 Ensuring the privacy and integrity of a hypervisor

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US14/741,147 Active 2033-05-15 US10140139B1 (en) 2012-06-19 2015-06-16 Ensuring the privacy and integrity of a hypervisor

Country Status (1)

Country Link
US (2) US10140139B1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US12265606B2 (en) * 2022-09-26 2025-04-01 Microsoft Technology Licensing, Llc Direct assignment of physical devices to confidential virtual machines

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10503771B2 (en) 2017-01-06 2019-12-10 Oracle International Corporation Efficient incremental backup and restoration of file system hierarchies with cloud object storage
US11429753B2 (en) * 2018-09-27 2022-08-30 Citrix Systems, Inc. Encryption of keyboard data to avoid being read by endpoint-hosted keylogger applications
US11977493B2 (en) * 2019-07-17 2024-05-07 Red Hat, Inc. Safe virtual machine physical device access for network function virtualization
US11467775B2 (en) 2019-10-15 2022-10-11 Hewlett Packard Enterprise Development Lp Virtual persistent volumes for containerized applications
US12135885B2 (en) 2020-01-31 2024-11-05 Hewlett Packard Enterprise Development Lp Virtual storage policies for virtual persistent volumes
US11687267B2 (en) 2020-04-14 2023-06-27 Hewlett Packard Enterprise Development Lp Containerized application manifests and virtual persistent volumes
US11693573B2 (en) 2020-06-18 2023-07-04 Hewlett Packard Enterprise Development Lp Relaying storage operation requests to storage systems using underlying volume identifiers
US11960773B2 (en) * 2020-07-31 2024-04-16 Hewlett Packard Enterprise Development Lp Modifying virtual persistent volumes based on analysis of performance metrics

Family Cites Families (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5657388A (en) * 1993-05-25 1997-08-12 Security Dynamics Technologies, Inc. Method and apparatus for utilizing a token for resource access
US6496847B1 (en) 1998-05-15 2002-12-17 Vmware, Inc. System and method for virtualizing computer systems
US6820142B2 (en) * 2000-12-14 2004-11-16 International Business Machines Corporation Token based DMA
US20030229794A1 (en) * 2002-06-07 2003-12-11 Sutton James A. System and method for protection against untrusted system management code by redirecting a system management interrupt and creating a virtual machine container
FR2849230B1 (en) 2002-12-24 2005-04-22 Francois Bangui METHOD AND APPARATUS FOR VERIFYING THE INTEGRITY OF A SOFTWARE APPLICATION WITHOUT AN ENCRYPTION / DECRYMENT KEY
US8055838B2 (en) 2003-07-22 2011-11-08 International Business Machines Corporation Apparatus and method for autonomically suspending and resuming logical partitions when I/O reconfiguration is required
US7467381B2 (en) 2003-12-16 2008-12-16 Intel Corporation Resource partitioning and direct access utilizing hardware support for virtualization
US7937700B1 (en) 2004-05-11 2011-05-03 Advanced Micro Devices, Inc. System, processor, and method for incremental state save/restore on world switch in a virtual machine environment
US8042163B1 (en) * 2004-05-20 2011-10-18 Symatec Operating Corporation Secure storage access using third party capability tokens
US7647589B1 (en) 2005-02-07 2010-01-12 Parallels Software International, Inc. Methods and systems for safe execution of guest code in virtual machine context
US8001342B2 (en) * 2006-03-29 2011-08-16 International Business Machines Corporation Method for storing and restoring persistent memory content and virtual machine state information
US8539137B1 (en) 2006-06-09 2013-09-17 Parallels IP Holdings GmbH System and method for management of virtual execution environment disk storage
US7636800B2 (en) 2006-06-27 2009-12-22 International Business Machines Corporation Method and system for memory address translation and pinning
US20080065854A1 (en) * 2006-09-07 2008-03-13 Sebastina Schoenberg Method and apparatus for accessing physical memory belonging to virtual machines from a user level monitor
US7761612B2 (en) 2006-12-07 2010-07-20 International Business Machines Corporation Migrating domains from one physical data processing system to another
US8276201B2 (en) * 2007-03-22 2012-09-25 International Business Machines Corporation Integrity protection in data processing systems
WO2009001153A1 (en) * 2007-06-28 2008-12-31 Nokia Corporation Memory protection unit in a virtual processing environment
US8819676B2 (en) 2007-10-30 2014-08-26 Vmware, Inc. Transparent memory-mapped emulation of I/O calls
US7743389B2 (en) 2007-11-06 2010-06-22 Vmware, Inc. Selecting between pass-through and emulation in a virtual machine environment
US8151032B2 (en) 2008-06-26 2012-04-03 Microsoft Corporation Direct memory access filter for virtualized operating systems
US20100332212A1 (en) 2008-09-19 2010-12-30 Ori Finkelman Method and apparatus for sleep and wake of computer devices
US20100077063A1 (en) 2008-09-19 2010-03-25 Jonathan Amit System and method for emulating a computing device
US8738932B2 (en) * 2009-01-16 2014-05-27 Teleputers, Llc System and method for processor-based security
US7904540B2 (en) 2009-03-24 2011-03-08 International Business Machines Corporation System and method for deploying virtual machines in a computing environment
US8555377B2 (en) 2010-04-29 2013-10-08 High Cloud Security Secure virtual machine
US8893306B2 (en) * 2010-08-31 2014-11-18 International Business Machines Corporation Resource management and security system
US8429322B2 (en) 2010-10-26 2013-04-23 Red Hat Israel, Ltd. Hotplug removal of a device in a virtual machine system
US8484392B2 (en) 2011-05-31 2013-07-09 Oracle International Corporation Method and system for infiniband host channel adaptor quality of service
US20130055254A1 (en) 2011-08-31 2013-02-28 Nokia Corporation Methods and apparatuses for providing a virtual machine with dynamic assignment of a physical hardware resource
US8549210B2 (en) 2011-09-20 2013-10-01 International Business Machines Corporation Mirroring virtual machines from a primary host to a secondary host
US8782351B2 (en) * 2011-10-13 2014-07-15 International Business Machines Corporation Protecting memory of a virtual guest
US9256552B2 (en) * 2011-11-21 2016-02-09 Cisco Technology, Inc. Selective access to executable memory
US9514507B2 (en) 2011-11-29 2016-12-06 Citrix Systems, Inc. Methods and systems for maintaining state in a virtual machine when disconnected from graphics hardware
US9361244B2 (en) * 2011-12-29 2016-06-07 Intel Corporation Apparatus for hardware accelerated runtime integrity measurement
US9021476B1 (en) 2012-06-19 2015-04-28 Bromium, Inc. Ensuring the privacy and integrity of a hypervisor

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US12265606B2 (en) * 2022-09-26 2025-04-01 Microsoft Technology Licensing, Llc Direct assignment of physical devices to confidential virtual machines

Also Published As

Publication number Publication date
US10140139B1 (en) 2018-11-27

Similar Documents

Publication Publication Date Title
US9021476B1 (en) Ensuring the privacy and integrity of a hypervisor
US20190065236A1 (en) Ensuring the privacy and integrity of a hypervisor
US11783081B2 (en) Secure public cloud
US11163911B2 (en) Secure public cloud with protected guest-verified host control
US11520611B2 (en) Secure public cloud using extended paging and memory integrity
US10686605B2 (en) Technologies for implementing mutually distrusting domains
US20250238255A1 (en) Scalable virtual machine operation inside trust domains within the trust domain architecture
JP6893479B2 (en) Data processing device and method using ownership table
US8776245B2 (en) Executing trusted applications with reduced trusted computing base
JP6913636B2 (en) Shared page
US20190147160A1 (en) Virtual machine manager facilitated selective code integrity enforcement
US8832352B2 (en) Hypervisor-driven protection of data from virtual machine clones
US10599565B2 (en) Hypervisor managing memory addressed above four gigabytes
US8893124B2 (en) Method, apparatus and system for limiting access to virtualization information in a memory
KR101323858B1 (en) Apparatus and method for controlling memory access in virtualized system
US9135046B1 (en) Preventing host operating system from inspecting or modifying data received by hardware controller by moving host operating system into a virtual machine after boot up
EP4020156B1 (en) Reducing latency of hardware trusted execution environments
US20150143484A1 (en) System and method for managing tokens authorizing on-device operations
US10395028B2 (en) Virtualization based intra-block workload isolation
US20250238380A1 (en) Method and apparatus to set guest physical address mapping attributes for trusted domain
KR20250020452A (en) Software isolation of virtual machine resources
US20240061697A1 (en) Providing trusted devices fine grained access into private memory of trusted execution environment

Legal Events

Date Code Title Description
AS Assignment

Owner name: BROMIUM, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PRATT, IAN;REEL/FRAME:047374/0333

Effective date: 20150610

STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BROMIUM, INC.;REEL/FRAME:051305/0894

Effective date: 20190919

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载