diff options
Diffstat (limited to 'ambari-server/src/main/resources/stacks/BigInsights/4.2/services/AMBARI_METRICS/configuration/ams-hbase-site.xml')
-rwxr-xr-x | ambari-server/src/main/resources/stacks/BigInsights/4.2/services/AMBARI_METRICS/configuration/ams-hbase-site.xml | 489 |
1 files changed, 0 insertions, 489 deletions
diff --git a/ambari-server/src/main/resources/stacks/BigInsights/4.2/services/AMBARI_METRICS/configuration/ams-hbase-site.xml b/ambari-server/src/main/resources/stacks/BigInsights/4.2/services/AMBARI_METRICS/configuration/ams-hbase-site.xml deleted file mode 100755 index bf62b8e365..0000000000 --- a/ambari-server/src/main/resources/stacks/BigInsights/4.2/services/AMBARI_METRICS/configuration/ams-hbase-site.xml +++ /dev/null @@ -1,489 +0,0 @@ -<?xml version="1.0"?> -<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> -<!-- -/** - * - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ ---> -<configuration> - <property> - <name>hbase.rootdir</name> - <value>file:///var/lib/ambari-metrics-collector/hbase</value> - <description> - Ambari Metrics service uses HBase as default storage backend. Set the rootdir for - HBase to either local filesystem path if using Ambari Metrics in embedded mode or - to a HDFS dir, example: hdfs://namenode.example.org:8020/amshbase. - </description> - <depends-on> - <property> - <type>core-site</type> - <name>fs.defaultFS</name> - </property> - <property> - <type>ams-site</type> - <name>timeline.metrics.service.operation.mode</name> - </property> - </depends-on> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.tmp.dir</name> - <value>/var/lib/ambari-metrics-collector/hbase-tmp</value> - <description> - Temporary directory on the local filesystem. - Change this setting to point to a location more permanent - than '/tmp' (The '/tmp' directory is often cleared on - machine restart). - </description> - <value-attributes> - <type>directory</type> - </value-attributes> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.local.dir</name> - <value>${hbase.tmp.dir}/local</value> - <description>Directory on the local filesystem to be used as a local storage - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.cluster.distributed</name> - <value>false</value> - <description> - The mode the cluster will be in. Possible values are false for - standalone mode and true for distributed mode. If false, startup will run - all HBase and ZooKeeper daemons together in the one JVM. - </description> - <depends-on> - <property> - <type>ams-site</type> - <name>timeline.metrics.service.operation.mode</name> - </property> - </depends-on> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.master.wait.on.regionservers.mintostart</name> - <value>1</value> - <description> - Ensure that HBase Master waits for # many region server to start. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.zookeeper.quorum</name> - <value>{{zookeeper_quorum_hosts}}</value> - <description>Comma separated list of servers in the ZooKeeper Quorum. - For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com". - By default this is set to localhost for local and pseudo-distributed modes - of operation. For a fully-distributed setup, this should be set to a full - list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh - this is the list of servers which we will start/stop ZooKeeper on. - </description> - <final>true</final> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.master.info.bindAddress</name> - <value>0.0.0.0</value> - <description>The bind address for the HBase Master web UI</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.master.info.port</name> - <value>61310</value> - <description>The port for the HBase Master web UI.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.regionserver.info.port</name> - <value>61330</value> - <description>The port for the HBase RegionServer web UI.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.master.port</name> - <value>61300</value> - <description>The port for the HBase Master web UI.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.regionserver.port</name> - <value>61320</value> - <description>The port for the HBase RegionServer web UI.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.hregion.majorcompaction</name> - <value>0</value> - <description> - The time (in milliseconds) between 'major' compactions of all - HStoreFiles in a region. - 0 to disable automated major compactions. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.query.spoolThresholdBytes</name> - <value>20971520</value> - <description> - Threshold size in bytes after which results from parallelly executed - query results are spooled to disk. Default is 20 mb. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.zookeeper.property.dataDir</name> - <value>${hbase.tmp.dir}/zookeeper</value> - <description> - Property from ZooKeeper's config zoo.cfg. - The directory where the snapshot is stored. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.client.scanner.caching</name> - <value>10000</value> - <description> - Number of rows that will be fetched when calling next on a scanner - if it is not served from (local, client) memory. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.normalizer.enabled</name> - <value>false</value> - <description>If set to true, Master will try to keep region size - within each table approximately the same.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.normalizer.period</name> - <value>600000</value> - <description>Period in ms at which the region normalizer runs in the Master.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.master.normalizer.class</name> - <value>org.apache.hadoop.hbase.master.normalizer.SimpleRegionNormalizer</value> - <description> - Class used to execute the region normalization when the period occurs. - See the class comment for more on how it works - http://hbase.apache.org/devapidocs/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.html - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hfile.block.cache.size</name> - <value>0.3</value> - <description> - Percentage of maximum heap (-Xmx setting) to allocate to block cache - used by a StoreFile. Default of 0.4 means allocate 40%. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.regionserver.global.memstore.upperLimit</name> - <value>0.5</value> - <description> - Maximum size of all memstores in a region server before new - updates are blocked and flushes are forced. Defaults to 40% of heap - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.regionserver.global.memstore.lowerLimit</name> - <value>0.4</value> - <description> - When memstores are being forced to flush to make room in - memory, keep flushing until we hit this mark. Defaults to 35% of heap. - This value equal to hbase.regionserver.global.memstore.upperLimit causes - the minimum possible flushing to occur when updates are blocked due to - memstore limiting. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.groupby.maxCacheSize</name> - <value>307200000</value> - <description> - Size in bytes of pages cached during GROUP BY spilling. Default is 100Mb. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.hregion.max.filesize</name> - <value>4294967296</value> - <description> - Maximum HFile size. If the sum of the sizes of a region’s HFiles has grown - to exceed this value, the region is split in two. Default is 10Gb. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.hregion.memstore.block.multiplier</name> - <value>4</value> - <description> - Block updates if memstore has hbase.hregion.memstore.block.multiplier - times hbase.hregion.memstore.flush.size bytes. Useful preventing runaway - memstore during spikes in update traffic. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.hstore.flusher.count</name> - <value>2</value> - <description> - The number of flush threads. With fewer threads, the MemStore flushes - will be queued. With more threads, the flushes will be executed in parallel, - increasing the load on HDFS, and potentially causing more compactions. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.query.timeoutMs</name> - <value>300000</value> - <description> - Number of milliseconds after which a query will timeout on the client. - Default is 5 min. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.query.keepAliveMs</name> - <value>300000</value> - <description> - Number of milliseconds after which a query will keep the connection to HBase alive. - Default is 5 min. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.client.scanner.timeout.period</name> - <value>300000</value> - <description> - Client scanner lease period in milliseconds. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.rpc.timeout</name> - <value>300000</value> - <description> - This is for the RPC layer to define how long HBase client applications - take for a remote call to time out. It uses pings to check connections - but will eventually throw a TimeoutException. - </description> - <display-name>HBase RPC Timeout</display-name> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.regionserver.thread.compaction.large</name> - <value>2</value> - <description> - Configuration key for the large compaction threads. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.regionserver.thread.compaction.small</name> - <value>3</value> - <description> - Configuration key for the small compaction threads. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.zookeeper.property.clientPort</name> - <value>{{zookeeper_clientPort}}</value> - <depends-on> - <property> - <type>zoo.cfg</type> - <name>clientPort</name> - </property> - <property> - <type>ams-hbase-site</type> - <name>hbase.cluster.distributed</name> - </property> - </depends-on> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.zookeeper.peerport</name> - <value>61288</value> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.zookeeper.leaderport</name> - <value>61388</value> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.hstore.blockingStoreFiles</name> - <value>200</value> - <description> - If more than this number of StoreFiles exist in any one Store - (one StoreFile is written per flush of MemStore), updates are blocked for - this region until a compaction is completed, or until - hbase.hstore.blockingWaitTime has been exceeded. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.hregion.memstore.flush.size</name> - <value>134217728</value> - <description> - Memstore will be flushed to disk if size of the memstore exceeds this - number of bytes. Value is checked by a thread that runs every - hbase.server.thread.wakefrequency. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.snapshot.enabled</name> - <value>false</value> - <description>Enable/Disable HBase snapshots.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.replication</name> - <value>false</value> - <description>Enable/Disable HBase replication.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>hbase.zookeeper.property.tickTime</name> - <value>6000</value> - <description> - The length of a single tick, which is the basic time unit used by - ZooKeeper, as measured in milliseconds. This property setting only - affects the ZK server started by AMS in embedded mode. Unit = ms. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>zookeeper.session.timeout</name> - <value>120000</value> - <description>ZooKeeper session timeout in milliseconds.</description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>zookeeper.session.timeout.localHBaseCluster</name> - <value>120000</value> - <description> - ZooKeeper session timeout in milliseconds for - pseudo distributed mode. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.sequence.saltBuckets</name> - <value>2</value> - <description> - Controls the number of pre-allocated regions for SYSTEM.SEQUENCE table. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.query.maxGlobalMemoryPercentage</name> - <value>15</value> - <description> - Percentage of total heap memory (i.e. Runtime.getRuntime().maxMemory()) - that all threads may use. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.spool.directory</name> - <value>${hbase.tmp.dir}/phoenix-spool</value> - <description> - Set directory for Phoenix spill files. If possible set this to a - different mount point from the one for hbase.rootdir in embedded mode. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.mutate.batchSize</name> - <value>10000</value> - <description> - The number of rows that are batched together and automatically committed - during the execution of an UPSERT SELECT or DELETE statement. - This affects performance of group by aggregators if they are being used. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.query.rowKeyOrderSaltedTable</name> - <value>true</value> - <description> - When set, we disallow user specified split points on salted table to ensure - that each bucket will only contains entries with the same salt byte. - When this property is turned on, the salted table would behave just like - a normal table and would return items in rowkey order for scans - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.coprocessor.maxServerCacheTimeToLiveMs</name> - <value>60000</value> - <description> - Maximum living time (in milliseconds) of server caches. A cache entry - expires after this amount of time has passed since last access. Consider - adjusting this parameter when a server-side IOException( - “Could not find hash cache for joinId”) happens. Getting warnings like - “Earlier hash cache(s) might have expired on servers” might also be a - sign that this number should be increased. - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>phoenix.coprocessor.maxMetaDataCacheSize</name> - <value>20480000</value> - <description> - Max size in bytes of total server-side metadata cache after which - evictions will begin to occur based on least recent access time. - Default is 20Mb - </description> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>dfs.client.read.shortcircuit</name> - <value>true</value> - <description>Enable/Disable short circuit read for your client. - Hadoop servers should be configured to allow short circuit read - for the hbase user for this to take effect - </description> - <depends-on> - <property> - <type>ams-hbase-site</type> - <name>hbase.cluster.distributed</name> - </property> - <property> - <type>ams-hbase-site</type> - <name>hbase.rootdir</name> - </property> - </depends-on> - <on-ambari-upgrade add="true"/> - </property> - <property> - <name>zookeeper.znode.parent</name> - <value>/ams-hbase-unsecure</value> - <on-ambari-upgrade add="true"/> - </property> -</configuration> |