diff --git a/gpdb-doc/dita/best_practices/platform.xml b/gpdb-doc/dita/best_practices/platform.xml deleted file mode 100644 index 78b9c39cd11cb235003ccbc6f3c55a4138b712f8..0000000000000000000000000000000000000000 --- a/gpdb-doc/dita/best_practices/platform.xml +++ /dev/null @@ -1,117 +0,0 @@ - - - - Hardware and Platform - -

In an MPP shared nothing environment, overall response time for a query is measured by the - completion time for all segments. Processing/query time is constrained by the slowest running - segment. While it is not enforced, it is highly recommended that all segment hosts have - identical configurations.

-
- CPU -

Choose a dual-socket PC with high clock rate, and multi-core CPUs for the Greenplum - Database master and segment hosts. If all of the segment hosts have the same or similar - configuration, it is easier to achieve uniform performance from the segments.

-
-
- Disk Storage -

Choose a hardware RAID storage system with 8 to 24 disks on the master and 16 to 24 disks - on the segment hosts. Controllers from LSI Corporation are used most frequently. SAS disks - generally provide better performance than SATA and are recommended.

-

Optimal I/O throughput is 2 GB per second read and 1 GB per second write.

-

RAID should be configured to allow for a hot spare disk to reduce the downtime in the event - of a single disk failure. RAID 5, RAID 6 and RAID 1 have all been used. A RAID mirroring - scheme should be selected to provide the least deterioration in performance during both disk - failure and RAID rebuilding scenarios.

-
-
- Network -

Networking throughput per host should be at least 10 gigabits per second, with 20 gigabits - per second preferred by using two 10-gigabit network interfaces on each host. If two - 10-gigabit networks are used, then configuring interface bonding can provide a single IP - address for each host used for sending and receiving data.

-

Disk patrolling and other periodically scheduled disk maintenance tasks should be either - disabled or the schedule should be controlled to prevent delays in the I/O subsystem at - unexpected times.

-
-
- Memory -

A minimum of 256GB per host is recommended.

-

Large amounts of RAM per host can be helpful if you plan to run queries with a high degree - of concurrency. Up to 1TB of RAM per host can improve performance with some workloads.

-

See , , , and for additional memory-related best - practices.

-
-
- Preferred Operating System -

Red Hat Enterprise Linux (RHEL) is the preferred operating system. See the Greenplum - Database Release Notes to find supported operating systems for a particular - release.

-
-
- File System -

XFS is the best practice file system for Greenplum Database data directories. XFS should be - mounted with the following mount options: - rw,noatime,inode64,allocsize=16m

-
-
- Port Configuration -

ip_local_port_range should be set up to not conflict with the Greenplum - Database port ranges. For example: - net.ipv4.ip_local_port_range = 3000 65535 -PORT_BASE=2000 -MIRROR_PORT_BASE=2100 -REPLICATION_PORT_BASE=2200 -MIRROR_REPLICATION_PORT_BASE=2300

-
-
- I/O Configuration -

blockdev read-ahead size should be set to 16384 on the devices that contain data - directories.

- /sbin/blockdev --getra /dev/sdb -16384 -

The deadline IO scheduler should be set for all data directory devices.

- # cat /sys/block/sdb/queue/scheduler - noop anticipatory [deadline] cfq -

The maximum number of OS files and processes should be increased in the - /etc/security/limits.conf file. - * soft nofile 65536 -* hard nofile 65536 -* soft nproc 131072 -* hard nproc 131072

-

Enable core files output to a known location and make sure limits.conf - allows core files. - kernel.core_pattern = /var/core/core.%h.%t -# grep core /etc/security/limits.conf -* soft core unlimited

-
-
- OS Memory Configuration -

The Linux sysctl vm.overcommit_memory and - vm.overcommit_ratio variables affect how the operating system manages - memory allocation. These variables should be set as follows:

-

vm.overcommit_memory determines the method the OS uses for determining how - much memory can be allocated to processes. This should be always set to 2, which is the only - safe setting for the database.

-

vm.overcommit_ratio is the percent of RAM that is used for application - processes. The default, 50, is the recommended setting.

-
-
- Shared Memory Settings -

Greenplum Database uses shared memory to communicate between postgres - processes that are part of the same postgres instance. The following shared - memory settings should be set in sysctl and are rarely modified.

- kernel.shmmax = 500000000 -kernel.shmmni = 4096 -kernel.shmall = 4000000000 -
-
- Validate the Operating System -

Run gpcheck (as root) to validate the operating system configuration. - gpcheck in the Greenplum Database Utility Guide.

-
- -