This is the source-only release of the Apache Trafodion (incubating) project. This release provides a number of new features and improvements across the project plus a large number of bug fixes.
Build instructions are available here.
Supported Platforms
The following platforms are supported in this release.
Operating Systems | CentOS 6.5 – 6.7 |
Hadoop Distributions | Cloudera CDH distribution 5.1 and Hortonworks distribution HDP 2.1 |
Java Version | JDK 1.7.0_67 or newer |
HBase Version | HBase 0.98.x |
Enhancements
This release contains the following new features.
Category | Feature | Defect ID |
---|---|---|
High Availability | Production Ready:
|
|
Performance | Production Ready:
|
|
Usability | Technology Preview:
|
|
Database Movement and Connectivity | Production Ready:
Technology Preview:
|
|
Manageability | Production Ready:
|
|
Security | Production Ready:
|
|
Stability |
|
Fixes
This release contains fixes to around 85 defects, including 25 critical defects, 54 high defects, and 10 medium and low defects. Those defects were filed through Launchpad.
Known Issues
All defect are located in Launchpad.
Getting TM error 97 when tables split or get moved
Defect: 1274651
Symptom: HBase Region Splits, Load Balancing, and Error 97.
Cause: As part of an HBase environment’s ongoing operations (and based on the policies configured for the HBase environment), an HBase region can either get split (into two daughter regions) or moved to a different region server. (Please see the blog: http://hortonworks.com/blog/apache-hbase-region-splitting-and-merging/.) If that happens when a Trafodion transaction is active (and operates on rows within the region that is being split or load-balanced), then a subsequent transaction commit operation by the application might encounter an error 97. Please note that under such conditions the Trafodion Transaction Manager will abort the transaction and will preserve the integrity of the database.
Solution: To minimize disruptions when this happens, we suggest that you use one or more of the following approaches:
- Enhance your JDBC application logic to retry when an error 97 is returned for a commit operation.
- Update the HBase configuration to reduce the times when such disruptions happen. It involves updates to some properties that can be set in hbase-site.xml (or via the manageability interface of your Hadoop distribution).
- Set the maximum file size of an HBase Region to 100 GB. For example, set the value of the property hbase.hregion.max.filesize to 107374182400.
-
Set the HBase region split policy to 'ConstantSizeRegionSplitPolicy'. For example, set the value of the property hbase.regionserver.region.split.policy to org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy.
NOTE: The split policy should already be set to 'onstantSizeRegionSplitPolicy' by the Trafodion installer.
Summary
Property Value hbase.hregion.max.filesize 107374182400 hbase.regionserver.region.split.policy org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy -
Disable HBase Region Load Balancing. Use the HBase shell command balance_switch false to disable the movement of a region from one server to another.
Example
hbase shell hbase(main):002:0> balance_switch false true -- Output will be the last setting of the balance_switch value 0 row(s) in 0.0080 seconds
- Pre-split the table into multiple regions by using the SALT USING n PARTITIONS clause when creating the table. The number of partitions that you specify could be a function of the number of region servers present in the HBase cluster. Here is a simple example in which the table INVENTORY is pre-split into four regions when created:
CREATE TABLE INVENTORY ( ITEM_ID INT UNSIGNED NO DEFAULT NOT NULL , ITEM_TYPE INT UNSIGNED NO DEFAULT NOT NULL , ITEM_COUNT INT UNSIGNED NO DEFAULT NOT NULL , PRIMARY KEY (ITEM_ID ASC) ) SALT USING 4 PARTITIONS ;
EXECUTE.BATCH update creates core-file
Defect: 1274962
Symptom: EXECUTE.BATCH hangs for a long time doing updates, and the update creates a core file.
Cause: To be determined.
Solution: Batch updates and ODBC row arrays do not currently work.
Random update statistics failures with HBase OutOfOrderScannerNextException
Defect: 1391271
Symptom: While running update statistics commands, you see HBase OutOfOrderScannerNextException errors.
Cause: The default hbase.rpc.timeout and hbase.client.scanner.timeout.period values might be too low given the size of the tables. Sampling in update statistics is implemented using the HBase Random RowFilter. For very large tables with several billion rows, the sampling ratio required to get a sample of 1 million rows is very small. This can result in HBase client connection timeout errors since there may be no row returned by a RegionServer for an extended period of time.
Solution: Increase the hbase.rpc.timeout and hbase.client.scanner.timeout.period values. We have found that increasing those values to 600 seconds (10 minutes) might sometimes prevent many timeout-related errors. For more information, see the HBase Configuration and Fine Tuning Recommendations.
If increasing the hbase.rpc.timeout and hbase.client.scanner.timeout.period values does not work, try increasing the chosen sampling size. Choose a sampling percentage higher than the default setting of 1 million rows for large tables. For example, suppose table T has one billion rows. The following UPDATE STATISTICS statement will sample a million rows, or approximately one-tenth of one percent of the total rows:
update statistics for table T on every column sample;
To sample one percent of the rows, regardless of the table size, you must explicitly state the sampling rate as follows:
update statistics for table T on every column sample random 1 percent;
Following update statistics, stats do not take effect immediately
Defect: 1409937
Symptom: Immediately following an update statistics operation, the generated query plan does not seem to reflect the existence of statistics. For example, in a session, you create, and populate a table and then run update statistics on the table, prepare a query, and exit. A serial plan is generated and the estimated cardinality is 100 for both tables. In a new session, you prepare the same query, and a parallel plan is generated where the estimated cardinality reflects the statistics.
Cause: This is a day-one issue.
Solution: Retry the query after two minutes. Set CQD HIST_NO_STATS_REFRESH_INTERVAL to ‘0’. Run an UPDATE STATISTICS statement. Perform DML operations in a different session.