MOT Memory and Storage Planning

This section describes the considerations and guidelines for evaluating, estimating and planning the quantity of memory and storage capacity to suit your specific application needs. This section also describes the various data aspects that affect the quantity of required memory, such as the size of data and indexes for the planned tables, memory to sustain transaction management and how fast the data is growing.

MOT Memory Planning

MOT belongs to the in-memory database class (IMDB) in which all tables and indexes reside entirely in memory.

NOTE: Memory storage is volatile, meaning that it requires power to maintain the stored information. Disk storage is persistent, meaning that it is written to disk, which is non-volatile storage. MOT uses both, having all data in memory, while persisting (by WAL logging) transactional changes to disk with strict consistency (in synchronous logging mode).

Sufficient physical memory must exist on the server in order to maintain the tables in their initial state, as well as to accommodate the related workload and growth of data. All this is in addition to the memory that is required for the traditional disk-based engine, tables and sessions that support the workload of disk-based tables. Therefore, planning ahead for enough memory to contain them all is essential.

Even so, you can get started with whatever amount of memory you have and perform basic tasks and evaluation tests. Later, when you are ready for production, the following issues should be addressed.

  • Memory Configuration Settings

    Similar to standard Postgres, the memory of the openGauss database process is controlled by the upper limit in its max_process_memory setting, which is defined in the **postgres.conf **file. The MOT engine and all its components and threads, reside within the openGauss process. Therefore, the memory allocated to MOT also operates within the upper boundary defined by max_process_memory for the entire openGauss database process.

    The amount of memory that MOT can reserve for itself is defined as a portion of max_process_memory. It can be defined either as a percentage or an absolute value that is less than max_process_memory. This portion is defined in the mot.conf configuration file by the <min / max>_mot_<global / local>_memory settings.

    The portion of max_process_memory that can be used by MOT must still leave at least 2 GB available for the Postgres (openGauss) envelope. Therefore, in order to ensure this, MOT verifies the following during database startup:

    (max_mot_global_memory + max_mot_local_memory) + 2GB < max_process_memory
    

    If this limit is breached, the internal MOT memory limits are adjusted to meet the preceding range to the maximum extent. This adjustment is performed during startup and calculates the value of MOT max memory accordingly.

    NOTE: MOT max memory is a logically calculated value of either the configured settings or their adjusted values of (max_mot_global_memory + max_mot_local_memory).

    In this case, a warning is issued to the server log, as shown below:

    The following is an example of a warning message reporting the problem:

    [WARNING] <Configuration> MOT engine maximum memory definitions (global: 9830 MB, local: 1843 MB, session large store: 0 MB, total: 11673 MB) breach GaussDB maximum process memory restriction (12288 MB) and/or total system memory (64243 MB). MOT values shall be adjusted accordingly to preserve required gap (2048 MB).  
    

    The following is an example of a warning message indicating that MOT is automatically adjusting the memory limits:

    [WARNING] <Configuration> Adjusting MOT memory limits: global = 8623 MB, local = 1617 MB, session large store = 0 MB, total = 10240 MB 
    

    This is the only place that shows the new memory limits.

    Additionally, MOT does not allow the insertion of additional data when the total memory usage approaches the chosen memory limits. The threshold for determining when additional data insertions are no longer allowed, is defined as a percentage of MOT max memory (which is a calculated value, as described above). The default is 90, meaning 90%. Attempting to add additional data over this threshold returns an error to the user and is also registered in the database log file.

  • Minimum and Maximum

    In order to secure memory for future operations, MOT pre-allocates memory based on the minimum global and local settings. The database administrator should specify the minimum amount of memory required for the MOT tables and sessions to sustain their workload. This ensures that this minimal memory is allocated to MOT even if another excessive memory-consuming application runs on the same server as the database and competes with the database for memory resources. The maximum values are used to limit memory growth.

  • Global and Local

    The memory used by MOT consists of two parts:

    • Global Memory: Global memory is a long-term memory pool that contains the data and indexes of MOT tables. It is evenly distributed across NUMA-nodes and is shared by all CPU cores.

    • **Local Memory **: Local memory is a memory pool used for short-term objects. Its primary consumers are sessions handling transactions. These sessions are storing data changes in the part of the memory dedicated to the relevant specific transaction (known as transaction private memory). Data changes are moved to the global memory at the commit phase. Memory object allocation is performed in NUMA-local manner in order to achieve the lowest possible latency.

      Deallocated objects are put back in the relevant memory pools. Minimal use of operating system memory allocation (malloc) functions during transactions circumvents unnecessary locks and latches.

      The allocation of these two memory parts is controlled by the dedicated min/max_mot_global_memory and min/max_mot_local_memory settings. If MOT global memory usage gets too close to this defined maximum, then MOT protects itself and does not accept new data. Attempts to allocate memory beyond this limit are denied and an error is reported to the user.

  • Minimum Memory Requirements

    To get started and perform a minimal evaluation of MOT performance, there are a few requirements.

    Make sure that the max_process_memory (as defined in postgres.conf) has sufficient capacity for MOT tables and sessions (configured by mix/max_mot_global_memory and mix/max_mot_local_memory), in addition to the disk tables buffer and extra memory. For simple tests, the default mot.conf settings can be used.

  • Actual Memory Requirements During Production

    In a typical OLTP workload, with 80:20 read:write ratio on average, MOT memory usage per table is 60% higher than in disk-based tables (this includes both the data and the indexes). This is due to the use of more optimal data structures and algorithms that enable faster access, with CPU-cache awareness and memory-prefetching.

    The actual memory requirement for a specific application depends on the quantity of data, the expected workload and especially on the data growth.

  • Max Global Memory Planning: Data + Index Size

    To plan for maximum global memory:

    1. Determine the size of a specific disk table (including both its data and all its indexes). The following statistical query can be used to determine the data size of the customer table and the customer_pkey index size:

      • Data size: select pg_relation_size('customer');
      • Index: select pg_relation_size('customer_pkey');
    2. Add 60%, which is the common requirement in MOT relative to the current size of the disk-based data and index.

    3. Add an additional percentage for the expected growth of data. For example:

      5% monthly growth = 80% yearly growth (1.05^12). Thus, in order to sustain a year's growth, allocate 80% more memory than is currently used by the tables.

      This completes the estimation and planning of the max_mot_global_memory value. The actual setting can be defined either as an absolute value or a percentage of the Postgres max_process_memory. The exact value is typically finetuned during deployment.

  • Max Local Memory Planning: Concurrent Session Support

    Local memory needs are primarily a function of the quantity of concurrent sessions. The typical OLTP workload of an average session uses up to 8 MB. This should be multiplied by the quantity of sessions and then a little bit extra should be added.

    A memory calculation can be performed in this manner and then finetuned, as follows:

    SESSION_COUNT * SESSION_SIZE (8 MB) + SOME_EXTRA (100MB should be enough)
    

    The default specifies 15% of Postgres's max_process_memory, which by default is 12 GB. This equals 1.8 GB, which is sufficient for 230 sessions, which is the requirement for the max_mot_local memory. The actual setting can be defined either as an absolute value or a percentage of the Postgres max_process_memory. The exact value is typically finetuned during deployment.

  • Unusually Large Transactions

    Some transactions are unusually large because they apply changes to a large number of rows. This may increase a single session's local memory up to the maximum allowed limit, which is 1 GB. For example:

    delete from SOME_VERY_LARGE_TABLE;
    

    Take this scenario into consideration when configuring the max_mot_local_memory setting, as well as during application development.

    NOTE: You may refer to Memory (MOT) for more information about configuration settings.

Storage I/O

MOT is a memory-optimized, persistent database storage engine. A disk drive(s) is required for storing the Redo Log (WAL) and a periodic checkpoint.

It is advised to use a storage device with low latency, such as SSD with a RAID-1 configuration, NVMe or any enterprise-grade storage system. When appropriate hardware is used, the database transaction processing and contention are the bottleneck, not the I/O.

Since the persistent storage is much slower than RAM memory, the I/O operations (logging and checkpoint) can create a bottleneck for both an in-memory and memory-optimized databases. However, MOT has a highly efficient durability design and implementation that is optimized for modern hardware (such as SSD and NVMe). In addition, MOT has minimized and optimized writing points (for example, by using parallel logging, a single log record per transaction and NUMA-aware transaction group writing) and has minimized the data written to disk (for example, only logging the delta or updated columns of the changed records and only logging a transaction at the commit phase).

Required Capacity

The required capacity is determined by the requirements of checkpoints and logging, as described below:

  • Checkpoint

    A checkpoint saves a snapshot of all the data to disk.

    Twice the size of all data should be allocated for checkpoints. There is no need to allocate space for the indexes for checkpoints.

    Checkpoint = 2 x MOT data size (rows only, index is not persistent)

    Twice the size is required because a snapshot is saved to disk of the entire size of the data, and in addition, the same amount of space should be allocated for the checkpoint that is in progress. When a checkpoint process finishes, the previous checkpoint files are deleted.

NOTE: In the next openGauss release, MOT will have an incremental checkpoint feature, which will significantly reduce this storage capacity requirement.

  • Logging

    MOT table log records are written to the same database transaction log as the other records of disk-based tables.

    The size of the log depends on the transactional throughput, the size of the data changes and the time between checkpoints (at each time checkpoint the Redo Log is truncated and starts to expand again).

    MOT tables use less log bandwidth and have lower I/O contention than disk-based tables. This is enabled by multiple mechanisms.

    For example, MOT does not log every operation before a transaction has been completed. It is only logged at the commit phase and only the updated delta record is logged (not full records like for disk-based tables).

    In order to ensure that the log I/O device does not become a bottleneck, the log file must be placed on a drive that has low latency.

NOTE: You may refer to Storage (MOT) for more information about configuration settings.

Feedback
编组 3备份
    openGauss 2024-04-22 00:47:24
    cancel