US20230153317A1 - Method for scheduling offloading snippets based on large amount of dbms task computation - Google Patents
Method for scheduling offloading snippets based on large amount of dbms task computation Download PDFInfo
- Publication number
- US20230153317A1 US20230153317A1 US17/985,994 US202217985994A US2023153317A1 US 20230153317 A1 US20230153317 A1 US 20230153317A1 US 202217985994 A US202217985994 A US 202217985994A US 2023153317 A1 US2023153317 A1 US 2023153317A1
- Authority
- US
- United States
- Prior art keywords
- query
- offloading
- dbms
- code
- computations
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
- G06F16/2453—Query optimisation
- G06F16/24534—Query rewriting; Transformation
- G06F16/24542—Plan optimisation
- G06F16/24545—Selectivity estimation or determination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/25—Integrating or interfacing systems involving database management systems
- G06F16/252—Integrating or interfacing systems involving database management systems between a Database Management System and a front-end application
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
- G06F16/2453—Query optimisation
- G06F16/24534—Query rewriting; Transformation
- G06F16/24542—Plan optimisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
- G06F16/2455—Query execution
- G06F16/24553—Query execution of query operations
- G06F16/24561—Intermediate data storage techniques for performance improvement
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/27—Replication, distribution or synchronisation of data between databases or within a distributed database system; Distributed database system architectures therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Program initiating; Program switching, e.g. by interrupt
- G06F9/4806—Task transfer initiation or dispatching
- G06F9/4843—Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
- G06F9/4881—Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues
Definitions
- the disclosure relates to a database (DB) processing technology, and more particularly, to a method for processing a query execution request of a client without a delay in an environment where a database management system (DBMS) and a storage in which a DB is established are disaggregated.
- DBMS database management system
- a database management system may acquire data by scanning a query in a database (DB), filter the acquired data, and return only the filtered data to a client.
- DBMS database management system
- an object of the present disclosure is to provide a method of defining a snippet for offloading a part of query computations in a DB system using a CSD, and scheduling the snippets which will be generated simultaneously for CSDs.
- a DB scheduling method may include: a step of receiving, by a DBMS, a query execution request from a clint; a step of determining, by the DBMS, whether to offload a part of query computations; a step of, when offloading is determined, generating, by the DBMS, an offloading code which is a code for offloading a part of the query computations, based on the received query; a first selection step of selecting, by the DBMS, one of a plurality of storages in which a DB is established; and a step of delivering, by the DBMS, the offloading code to the storage selected at the first selection step.
- the step of determining may include determining by determining which is appropriate, the DBMS processing all of the query computations, or the storage processing a part of the query computations and the DMBS processing the other query computations.
- the step of determining may include determining to offload a part of the query computations when there is a large amount of data transmitted between the DBMS and the storage.
- the first selection step may include selecting a storage that has the most unused resources by monitoring a resource state of the storages.
- the second selection step may include selecting a CSD that has most unused resources by monitoring a resource state of the CSDs.
- the offloading code may be generated as a snippet, and may include a parsed query and BufferAddress, and the BufferAddress may be an address of a buffer which stores a result of performing a part of the query computations by the offloading code.
- a DB system may include: a plurality of storages in which a DB is established; and a DBMS configured to determine whether to offload a part of query computations upon receiving a query execution request from a client, to generate an offloading code which is a code for offloading a part of the query computations, based on the received query, when offloading is determined, to select one of the plurality of storages, and to deliver the offloading code.
- a DB system may include a plurality of storages in which a DB is established, and a DBMS configured to generate an offloading code which is a code for offloading query computations, based on a query received from a client, to select one of a plurality of storages, and to deliver the offloading code.
- a snippet for offloading a part of query computations may be defined, and snippets which will be generated simultaneously are scheduled for CSDs, so that CSD resources are equally utilized, a query execution time is reduced, and reliability on data processing is enhanced.
- FIG. 1 is a view illustrating a DB system according to an embodiment of the disclosure
- FIG. 2 is a view illustrating a structure of the DBMS shown in FIG. 1 ;
- FIG. 3 is a view illustrating a structure of a query computation offloading snippet
- FIG. 4 is a view provided to explain a process of delivering a query computation offloading snippet
- FIG. 5 is a view illustrating a structure of the storage node shown in FIG. 1 ;
- An embodiment of the disclosure provides a method for defining a snippet for offloading (pushing down) a part of query computations onto a computational storage drive (CSD) in a DB system using the CSD, and scheduling the snippet by considering a resource state of the CSD.
- CSD computational storage drive
- FIG. 1 is a view illustrating a DB system according to an embodiment of the disclosure.
- the DB system according to an embodiment of the disclosure may be divided into a database management system (DBMS) 100 and a plurality of storage nodes 200 - 1 , 200 - 2 , ..., 200 - n as shown in the drawing.
- DBMS database management system
- the storge nodes 200 - 1 , 200 - 2 , ..., 200 - n refer to storage systems in which the same DB is established.
- the plurality of storage nodes 200 - 1 , 200 - 2 , ..., 200 - n may be implemented in order to process many queries which are requested simultaneously.
- the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n include a computation function, and may perform a part of query computations.
- the DBMS 100 is a system that executes a query according to a request of a client (not shown) and returns a result thereof. All of the query computations may be performed by the DBMS 100 , but a part of the query computations may be offloaded onto the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n .
- FIG. 2 is a view illustrating a structure of the DBMS 100 shown in FIG. 1 .
- the DBMS 100 may include a query engine 110 , a storage engine 120 , and a node scheduler 130 .
- the query engine 110 may optimize the received query and deliver the same to the storage engine 120 .
- the storage engine 120 may determine whether to offload query computations. Offloading query computations may refer to offloading a part of the query computations onto the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n . A part of the query computations may include query scanning, filtering, and validation.
- the storage engine 120 generates a query computation offloading snippet based on the query, and delivers the generated query computation offloading snippet to the node scheduler 130 .
- the query computation offloading snippet generated by the storage engine 120 is a source code for offloading a part of the query computations onto a storage node, and a detailed structure thereof is illustrated in FIG. 3 .
- the query computation offloading snippet includes ParsedQuery, TableSchema, BlockOffset, and BufferAddress.
- the BufferAddress refers to an address of a buffer (not shown) that stores a result of performing a part of the query computations according to the query computation offloading snippet.
- the buffer is provided in the storage engine 120 .
- the node scheduler 130 delivers the query computation offloading snippet generated by the storage engine 120 to the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n . This process is illustrated in FIG. 4 .
- a node data manager of the node scheduler 130 may identify a DB table that should perform a query computation offloading snippet (2), and an API manager of the node scheduler 130 requests a scheduler manager to schedule the query computation offloading snippet (3).
- the scheduler manager of the node scheduler 130 may select one of the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n , and may deliver the query computation offloading snippet (4).
- FIG. 5 is a view illustrating a structure of the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n shown in FIG. 1 . Since the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n may be implemented in the same way, only one storage node 200 is illustrated in FIG. 5 as a representative by using reference numeral “200.”
- the storage node 200 may include a storage I/O interface 210 , a CSD scheduler 220 , and a plurality of CSDs 230 - 1 , 230 - 2 , 230 - 3 .
- the storage I/O interface 210 may be an interface for communicating with the DBMS 100 .
- the CSDs 230 - 1 , 230 - 2 , 230 - 3 include not only storage resources for storing a DB, but also CPU resources (for example, a CSD runtime) for performing a part of query computations.
- the CSD scheduler 220 may select one of the CSDs 230 - 1 , 230 - 3 , 230 - 3 and may deliver the query computation offloading snippet.
- FIG. 6 is a view provided to explain a query computation offloading snippet scheduling method according to another embodiment of the disclosure.
- an optimizer of the query engine 110 may optimize the received query (2) and may deliver the query to the storage engine 120 (3).
- a table manager of the storage engine 120 may determine whether to offload query computations, and, when offloading is determined, the table manager may generate a query computation offloading snippet based on the query (4).
- offloading the query computations may be appropriate, but in other cases, offloading the query computations may be inappropriate.
- the node scheduler 130 may select one of the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n , and may deliver the query computation offloading snippet (5).
- the node scheduler 130 may monitor a resource state (a CPU core, a memory) of the storage nodes 200 - 1 , 200 - 2 , ..., 200 - n , and may select a storage node that has the most unused resources. This is for equal utilization of resources, reduction of a query computation time, and enhancement of data processing reliability.
- a resource state a CPU core, a memory
- the CSD scheduler 220 of the storage node which receives the query computation offloading snippet may select one of the plurality of CSDs 230 - 1 , 230 - 2 , 230 - 3 owned by the CSD scheduler 220 , and may deliver the query computation offloading snippet (6).
- the CSD scheduler 220 may monitor a resource state (a CPU core, a memory) of the CSDs 230 - 1 , 230 - 2 , 230 - 3 , and may select a CSD that has the most unused resources. This is for equal utilization of resources, reduction of a query computation time, and enhancement of data processing reliability.
- a resource state a CPU core, a memory
- the selected CSD may receive the query computation offloading snippet, and may perform a part of the query computations, and return a result thereof to the buffer of the storage engine 120 .
- An address of the buffer to which the result is returned may be indicated by the query computation offloading snippet.
- a query computation offloading snippet is proposed as a structure for offloading a part of query computations from a DMBS onto a storage when the storage is implemented by using a CSD in a cloud-native environment, in which the DBMS and the storage are disaggregated.
- computation offloading snippets which are generated simultaneously due to a large amount of queries are scheduled based on a resource state of a CSD, so that resources are equally utilized, a response time is reduced, and reliability on data processing is enhanced.
- the technical concept of the disclosure may be applied to a computer-readable recording medium which records a computer program for performing the functions of the apparatus and the method according to the present embodiments.
- the technical idea according to various embodiments of the present disclosure may be implemented in the form of a computer readable code recorded on the computer-readable recording medium.
- the computer-readable recording medium may be any data storage device that can be read by a computer and can store data.
- the computer-readable recording medium may be a read only memory (ROM), a random access memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical disk, a hard disk drive, or the like.
- a computer readable code or program that is stored in the computer readable recording medium may be transmitted via a network connected between computers.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- Operations Research (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
- This application is based on and claims priority under 35 U.S.C. § 119 to Korean Patent Application No. 10-2021-0156264, filed on Nov. 15, 2021, in the Korean Intellectual Property Office, the disclosure of which is herein incorporated by reference in its entirety.
- The disclosure relates to a database (DB) processing technology, and more particularly, to a method for processing a query execution request of a client without a delay in an environment where a database management system (DBMS) and a storage in which a DB is established are disaggregated.
- In a query execution process, a database management system (DBMS) may acquire data by scanning a query in a database (DB), filter the acquired data, and return only the filtered data to a client.
- However, when there are many data to be scanned in the DB, many data are transmitted from a storage in which the DB is established to the DBMS, and accordingly, there are problems that a bandwidth becomes scarce and an overall response speed is reduced, and power consumption increases.
- As a solution to these problems, there is a method of implementing the storage of the DB by using a computational storage drive (CSD), and allowing the storage to perform a part of the query computations. However, a specific method to achieve this has not yet been proposed.
- The disclosure has been developed to address the above-discussed deficiencies of the prior art, and an object of the present disclosure is to provide a method of defining a snippet for offloading a part of query computations in a DB system using a CSD, and scheduling the snippets which will be generated simultaneously for CSDs.
- According to an embodiment of the disclosure to achieve the above-described object, a DB scheduling method may include: a step of receiving, by a DBMS, a query execution request from a clint; a step of determining, by the DBMS, whether to offload a part of query computations; a step of, when offloading is determined, generating, by the DBMS, an offloading code which is a code for offloading a part of the query computations, based on the received query; a first selection step of selecting, by the DBMS, one of a plurality of storages in which a DB is established; and a step of delivering, by the DBMS, the offloading code to the storage selected at the first selection step.
- The step of determining may include determining by determining which is appropriate, the DBMS processing all of the query computations, or the storage processing a part of the query computations and the DMBS processing the other query computations.
- The step of determining may include determining to offload a part of the query computations when there is a large amount of data transmitted between the DBMS and the storage.
- The first selection step may include selecting a storage that has the most unused resources by monitoring a resource state of the storages.
- According to an embodiment of the disclosure, the DB scheduling method may further include: a second selection step of selecting, by the storage which receives the offloading code, one of a plurality of CDSs; and a step of delivering, by the storage, the received offloading code to the CSD selected at the second selection step.
- The second selection step may include selecting a CSD that has most unused resources by monitoring a resource state of the CSDs.
- The offloading code may be generated as a snippet, and may include a parsed query and BufferAddress, and the BufferAddress may be an address of a buffer which stores a result of performing a part of the query computations by the offloading code.
- The buffer may be provided in the DBMS. The DBMS and the storages may be disaggregated.
- According to another embodiment of the disclosure, a DB system may include: a plurality of storages in which a DB is established; and a DBMS configured to determine whether to offload a part of query computations upon receiving a query execution request from a client, to generate an offloading code which is a code for offloading a part of the query computations, based on the received query, when offloading is determined, to select one of the plurality of storages, and to deliver the offloading code.
- According to still another embodiment of the disclosure, a DB scheduling method may include: a step of generating, by a DBMS, an offloading code which is a code for offloading query computations, based on a query received from a client; a first selection step of selecting, by the DBMS, one of a plurality of storages in which a DB is established; and a step of delivering, by the DBMS, the offloading code to the storage selected at the first selection step.
- According to yet another embodiment of the disclosure, a DB system may include a plurality of storages in which a DB is established, and a DBMS configured to generate an offloading code which is a code for offloading query computations, based on a query received from a client, to select one of a plurality of storages, and to deliver the offloading code.
- According to embodiments of the disclosure as described above, in a DB system using a CSD, a snippet for offloading a part of query computations may be defined, and snippets which will be generated simultaneously are scheduled for CSDs, so that CSD resources are equally utilized, a query execution time is reduced, and reliability on data processing is enhanced.
- Other aspects, advantages, and salient features of the invention will become apparent to those skilled in the art from the following detailed description, which, taken in conjunction with the annexed drawings, discloses exemplary embodiments of the invention.
- Before undertaking the DETAILED DESCRIPTION OF THE INVENTION below, it may be advantageous to set forth definitions of certain words and phrases used throughout this patent document: the terms “include” and “comprise,” as well as derivatives thereof, mean inclusion without limitation; the term “or,” is inclusive, meaning and/or; the phrases “associated with” and “associated therewith,” as well as derivatives thereof, may mean to include, be included within, interconnect with, contain, be contained within, connect to or with, couple to or with, be communicable with, cooperate with, interleave, juxtapose, be proximate to, be bound to or with, have, have a property of, or the like. Definitions for certain words and phrases are provided throughout this patent document, those of ordinary skill in the art should understand that in many, if not most instances, such definitions apply to prior, as well as future uses of such defined words and phrases.
- For a more complete understanding of the present disclosure and its advantages, reference is now made to the following description taken in conjunction with the accompanying drawings, in which like reference numerals represent like parts:
-
FIG. 1 is a view illustrating a DB system according to an embodiment of the disclosure; -
FIG. 2 is a view illustrating a structure of the DBMS shown inFIG. 1 ; -
FIG. 3 is a view illustrating a structure of a query computation offloading snippet; -
FIG. 4 is a view provided to explain a process of delivering a query computation offloading snippet; -
FIG. 5 is a view illustrating a structure of the storage node shown inFIG. 1 ; and -
FIG. 6 is a view provided to explain a query computation offloading snippet scheduling method according to another embodiment of the disclosure. - Hereinafter, the disclosure will be described in detail with reference to the accompanying drawings.
- An embodiment of the disclosure provides a method for defining a snippet for offloading (pushing down) a part of query computations onto a computational storage drive (CSD) in a DB system using the CSD, and scheduling the snippet by considering a resource state of the CSD.
-
FIG. 1 is a view illustrating a DB system according to an embodiment of the disclosure. The DB system according to an embodiment of the disclosure may be divided into a database management system (DBMS) 100 and a plurality of storage nodes 200-1, 200-2, ..., 200-n as shown in the drawing. - The storge nodes 200-1, 200-2, ..., 200-n refer to storage systems in which the same DB is established. The plurality of storage nodes 200-1, 200-2, ..., 200-n may be implemented in order to process many queries which are requested simultaneously.
- Furthermore, the storage nodes 200-1, 200-2, ..., 200-n include a computation function, and may perform a part of query computations.
- The DBMS 100 is a system that executes a query according to a request of a client (not shown) and returns a result thereof. All of the query computations may be performed by the DBMS 100, but a part of the query computations may be offloaded onto the storage nodes 200-1, 200-2, ..., 200-n.
-
FIG. 2 is a view illustrating a structure of theDBMS 100 shown inFIG. 1 . As shown in the drawing, the DBMS 100 may include aquery engine 110, astorage engine 120, and anode scheduler 130. - Upon receiving a query execution request from a client, the
query engine 110 may optimize the received query and deliver the same to thestorage engine 120. - The
storage engine 120 may determine whether to offload query computations. Offloading query computations may refer to offloading a part of the query computations onto the storage nodes 200-1, 200-2, ..., 200-n. A part of the query computations may include query scanning, filtering, and validation. - The
storage engine 120 generates a query computation offloading snippet based on the query, and delivers the generated query computation offloading snippet to thenode scheduler 130. - The query computation offloading snippet generated by the
storage engine 120 is a source code for offloading a part of the query computations onto a storage node, and a detailed structure thereof is illustrated inFIG. 3 . - As shown in the drawing, the query computation offloading snippet includes ParsedQuery, TableSchema, BlockOffset, and BufferAddress. The BufferAddress refers to an address of a buffer (not shown) that stores a result of performing a part of the query computations according to the query computation offloading snippet. The buffer is provided in the
storage engine 120. - The
node scheduler 130 delivers the query computation offloading snippet generated by thestorage engine 120 to the storage nodes 200-1, 200-2, ..., 200-n. This process is illustrated inFIG. 4 . - As shown in the drawing, when the query computation offloading snippet is inputted from the
storage engine 120, a snippet validator of thenode scheduler 130 checks validity of the query computation offloading snippet (1). - In addition, a node data manager of the
node scheduler 130 may identify a DB table that should perform a query computation offloading snippet (2), and an API manager of thenode scheduler 130 requests a scheduler manager to schedule the query computation offloading snippet (3). - Accordingly, the scheduler manager of the
node scheduler 130 may select one of the storage nodes 200-1, 200-2, ..., 200-n, and may deliver the query computation offloading snippet (4). -
FIG. 5 is a view illustrating a structure of the storage nodes 200-1, 200-2, ..., 200-n shown inFIG. 1 . Since the storage nodes 200-1, 200-2, ..., 200-n may be implemented in the same way, only onestorage node 200 is illustrated inFIG. 5 as a representative by using reference numeral “200.” - As shown in the drawing, the
storage node 200 may include a storage I/O interface 210, aCSD scheduler 220, and a plurality of CSDs 230-1, 230-2, 230-3. - The storage I/
O interface 210 may be an interface for communicating with theDBMS 100. The CSDs 230-1, 230-2, 230-3 include not only storage resources for storing a DB, but also CPU resources (for example, a CSD runtime) for performing a part of query computations. - When a query computation offloading snippet is delivered from the
node scheduler 130 of theDBMS 100 through the storage I/O interface 210, theCSD scheduler 220 may select one of the CSDs 230-1, 230-3, 230-3 and may deliver the query computation offloading snippet. -
FIG. 6 is a view provided to explain a query computation offloading snippet scheduling method according to another embodiment of the disclosure. - When the
query engine 110 of theDBMS 100 receives a query execution request from a client (not shown) (1), an optimizer of thequery engine 110 may optimize the received query (2) and may deliver the query to the storage engine 120 (3). - Then, a table manager of the
storage engine 120 may determine whether to offload query computations, and, when offloading is determined, the table manager may generate a query computation offloading snippet based on the query (4). - It may be determined whether the query computations are offloaded by determining whether it is appropriate for the
query engine 110 to execute all of the query computations, or whether it is appropriate for the storage nodes 200-1, 200-2, ..., 200-n to execute a part of the query computations, and for thequery engine 110 to execute the other query computations. - When there is a large amount of data transmitted between the
DBMS 100 and the storage nodes 200-1, 200-2, ..., 200-n in executing a query, that is, when there are many data to be scanned, offloading the query computations may be appropriate, but in other cases, offloading the query computations may be inappropriate. - Herein, it is assumed that, in operation “④,” offloading the query computations is determined and the query computation offloading snippet is generated.
- The
node scheduler 130 may select one of the storage nodes 200-1, 200-2, ..., 200-n, and may deliver the query computation offloading snippet (5). - Herein, the
node scheduler 130 may monitor a resource state (a CPU core, a memory) of the storage nodes 200-1, 200-2, ..., 200-n, and may select a storage node that has the most unused resources. This is for equal utilization of resources, reduction of a query computation time, and enhancement of data processing reliability. - The
CSD scheduler 220 of the storage node which receives the query computation offloading snippet may select one of the plurality of CSDs 230-1, 230-2, 230-3 owned by theCSD scheduler 220, and may deliver the query computation offloading snippet (6). - Herein, the
CSD scheduler 220 may monitor a resource state (a CPU core, a memory) of the CSDs 230-1, 230-2, 230-3, and may select a CSD that has the most unused resources. This is for equal utilization of resources, reduction of a query computation time, and enhancement of data processing reliability. - Thereafter, the selected CSD may receive the query computation offloading snippet, and may perform a part of the query computations, and return a result thereof to the buffer of the
storage engine 120. An address of the buffer to which the result is returned may be indicated by the query computation offloading snippet. - Up to now, a method for defining a snippet for offloading a part of query computations to a CSD in a DB system using the CSD, and scheduling the snippet by considering a resource state of the CSD have been described in detail with reference to preferred embodiments.
- Specifically, in an embodiment of the disclosure, a query computation offloading snippet is proposed as a structure for offloading a part of query computations from a DMBS onto a storage when the storage is implemented by using a CSD in a cloud-native environment, in which the DBMS and the storage are disaggregated.
- In addition, in an embodiment of the disclosure, computation offloading snippets which are generated simultaneously due to a large amount of queries are scheduled based on a resource state of a CSD, so that resources are equally utilized, a response time is reduced, and reliability on data processing is enhanced.
- The technical concept of the disclosure may be applied to a computer-readable recording medium which records a computer program for performing the functions of the apparatus and the method according to the present embodiments. In addition, the technical idea according to various embodiments of the present disclosure may be implemented in the form of a computer readable code recorded on the computer-readable recording medium. The computer-readable recording medium may be any data storage device that can be read by a computer and can store data. For example, the computer-readable recording medium may be a read only memory (ROM), a random access memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical disk, a hard disk drive, or the like. A computer readable code or program that is stored in the computer readable recording medium may be transmitted via a network connected between computers.
- In addition, while preferred embodiments of the present disclosure have been illustrated and described, the present disclosure is not limited to the above-described specific embodiments. Various changes can be made by a person skilled in the art without departing from the scope of the present disclosure claimed in claims, and also, changed embodiments should not be understood as being separate from the technical idea or prospect of the present disclosure.
Claims (11)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020210156264A KR102628592B1 (en) | 2021-11-15 | 2021-11-15 | Method for scheduling offloading snippets based on large amount of DBMS task computation |
KR10-2021-0156264 | 2021-11-15 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20230153317A1 true US20230153317A1 (en) | 2023-05-18 |
Family
ID=86323567
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/985,994 Pending US20230153317A1 (en) | 2021-11-15 | 2022-11-14 | Method for scheduling offloading snippets based on large amount of dbms task computation |
Country Status (2)
Country | Link |
---|---|
US (1) | US20230153317A1 (en) |
KR (1) | KR102628592B1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20230153306A1 (en) * | 2021-11-15 | 2023-05-18 | Korea Electronics Technology Institute | Offloading data interfacing method between dbms storage engine and computational storage device |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080183688A1 (en) * | 2006-08-25 | 2008-07-31 | Chamdani Joseph I | Methods and systems for hardware acceleration of database operations and queries |
US7885953B2 (en) * | 2007-10-03 | 2011-02-08 | International Business Machines Corporation | Off-loading star join operations to a storage server |
US20140207755A1 (en) * | 2012-11-19 | 2014-07-24 | Nec Laboratories America, Inc. | System for Multi-store Analytics Execution Environments with Storage Constraints |
US20180349458A1 (en) * | 2017-05-31 | 2018-12-06 | Oracle International Corporation | Consistent query execution for big data analytics in a hybrid database |
US20200394191A1 (en) * | 2019-06-14 | 2020-12-17 | Oracle International Corporation | Non-disruptive dynamic ad-hoc database catalog services |
US20210073232A1 (en) * | 2019-09-11 | 2021-03-11 | Sap Se | Database management system query plan cache management |
US20210271680A1 (en) * | 2020-02-27 | 2021-09-02 | Sap Se | Near-memory acceleration for database operations |
US20220156163A1 (en) * | 2020-11-13 | 2022-05-19 | Oracle International Corporation | Fault tolerance in scale-out distributed query processing appliance |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102154446B1 (en) * | 2019-11-14 | 2020-09-09 | 한국전자기술연구원 | Method for fast scheduling for resource balanced allocation on distributed and collaborative container platform environment |
-
2021
- 2021-11-15 KR KR1020210156264A patent/KR102628592B1/en active Active
-
2022
- 2022-11-14 US US17/985,994 patent/US20230153317A1/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080183688A1 (en) * | 2006-08-25 | 2008-07-31 | Chamdani Joseph I | Methods and systems for hardware acceleration of database operations and queries |
US7885953B2 (en) * | 2007-10-03 | 2011-02-08 | International Business Machines Corporation | Off-loading star join operations to a storage server |
US20140207755A1 (en) * | 2012-11-19 | 2014-07-24 | Nec Laboratories America, Inc. | System for Multi-store Analytics Execution Environments with Storage Constraints |
US20180349458A1 (en) * | 2017-05-31 | 2018-12-06 | Oracle International Corporation | Consistent query execution for big data analytics in a hybrid database |
US20200394191A1 (en) * | 2019-06-14 | 2020-12-17 | Oracle International Corporation | Non-disruptive dynamic ad-hoc database catalog services |
US20210073232A1 (en) * | 2019-09-11 | 2021-03-11 | Sap Se | Database management system query plan cache management |
US20210271680A1 (en) * | 2020-02-27 | 2021-09-02 | Sap Se | Near-memory acceleration for database operations |
US20220156163A1 (en) * | 2020-11-13 | 2022-05-19 | Oracle International Corporation | Fault tolerance in scale-out distributed query processing appliance |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20230153306A1 (en) * | 2021-11-15 | 2023-05-18 | Korea Electronics Technology Institute | Offloading data interfacing method between dbms storage engine and computational storage device |
Also Published As
Publication number | Publication date |
---|---|
KR102628592B1 (en) | 2024-01-25 |
KR20230070567A (en) | 2023-05-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3545435B1 (en) | Database system and method for compiling serial and parallel database query execution plans | |
US11422853B2 (en) | Dynamic tree determination for data processing | |
US20200349160A1 (en) | Data query method, apparatus and device | |
US9436739B2 (en) | Dynamic priority-based query scheduling | |
US20230124520A1 (en) | Task execution method and storage device | |
US8135703B2 (en) | Multi-partition query governor in a computer database system | |
US20190253489A1 (en) | Command process load balancing system | |
US10261888B2 (en) | Emulating an environment of a target database system | |
US20170228422A1 (en) | Flexible task scheduler for multiple parallel processing of database data | |
US7792819B2 (en) | Priority reduction for fast partitions during query execution | |
US7624208B2 (en) | Method, system, and computer program for managing a queuing system | |
US20150112934A1 (en) | Parallel scanners for log based replication | |
CN111309805B (en) | Data reading and writing method and device for database | |
US20230153317A1 (en) | Method for scheduling offloading snippets based on large amount of dbms task computation | |
US5761696A (en) | Parallel database serving mechanism for a single-level-store computer system | |
CN106648874B (en) | Processing method and device for batch tasks | |
US7933948B2 (en) | Computer-readable medium to multiplex multiple application server requests over a single database connection | |
US11281654B2 (en) | Customized roll back strategy for databases in mixed workload environments | |
US20130006955A1 (en) | Method and system for processing data for database modification | |
CN118535546A (en) | Database performance optimization method and device | |
US20160335321A1 (en) | Database management system, computer, and database management method | |
US20230153306A1 (en) | Offloading data interfacing method between dbms storage engine and computational storage device | |
US7721287B2 (en) | Organizing transmission of repository data | |
US20240160610A1 (en) | Method of applying computational model in multiple computational storage-based dbms storage engine | |
CN114070820A (en) | Domain name redirection method, device, medium and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KOREA ELECTRONICS TECHNOLOGY INSTITUTE, KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AN, JAE HOON;KIM, YOUNG HWAN;REEL/FRAME:061754/0473 Effective date: 20221108 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |