Skip to content

Commit

Permalink
HBASE-25601 Use ASF-official mailing list archives
Browse files Browse the repository at this point in the history
Signed-off-by: Peter Somogyi <[email protected]>
Signed-off-by: Duo Zhang <[email protected]>

Closes apache#2983
  • Loading branch information
joshelser authored and dmisen committed Apr 8, 2022
1 parent 29f6eb1 commit e1c9299
Show file tree
Hide file tree
Showing 9 changed files with 29 additions and 42 deletions.
2 changes: 0 additions & 2 deletions pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -115,7 +115,6 @@
<archive>https://mail-archives.apache.org/mod_mbox/hbase-user/</archive>
<otherArchives>
<otherArchive>https://dir.gmane.org/gmane.comp.java.hadoop.hbase.user</otherArchive>
<otherArchive>https://search-hadoop.com/?q=&amp;fc_project=HBase</otherArchive>
</otherArchives>
</mailingList>
<mailingList>
Expand All @@ -126,7 +125,6 @@
<archive>https://mail-archives.apache.org/mod_mbox/hbase-dev/</archive>
<otherArchives>
<otherArchive>https://dir.gmane.org/gmane.comp.java.hadoop.hbase.devel</otherArchive>
<otherArchive>https://search-hadoop.com/?q=&amp;fc_project=HBase</otherArchive>
</otherArchives>
</mailingList>
<mailingList>
Expand Down
8 changes: 4 additions & 4 deletions src/main/asciidoc/_chapters/community.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ Just request the name of your branch be added to JIRA up on the developer's mail
Thereafter you can file issues against your feature branch in Apache HBase JIRA.
Your code you keep elsewhere -- it should be public so it can be observed -- and you can update dev mailing list on progress.
When the feature is ready for commit, 3 +1s from committers will get your feature merged.
See link:http://search-hadoop.com/m/asM982C5FkS1[HBase, mail # dev - Thoughts
See link:https://lists.apache.org/thread.html/200513c7e7e4df23c8b9134eeee009d61205c79314e77f222d396006%401346870308%40%3Cdev.hbase.apache.org%3E[HBase, mail # dev - Thoughts
about large feature dev branches]
[[patchplusonepolicy]]
Expand All @@ -61,7 +61,7 @@ Any -1 on a patch by anyone vetos a patch; it cannot be committed until the just
[[hbase.fix.version.in.jira]]
.How to set fix version in JIRA on issue resolve
Here is how link:http://search-hadoop.com/m/azemIi5RCJ1[we agreed] to set versions in JIRA when we resolve an issue.
Here is how we agreed to set versions in JIRA when we resolve an issue.
If trunk is going to be 0.98.0 then:
* Commit only to trunk: Mark with 0.98
Expand All @@ -73,7 +73,7 @@ If trunk is going to be 0.98.0 then:
[[hbase.when.to.close.jira]]
.Policy on when to set a RESOLVED JIRA as CLOSED
We link:http://search-hadoop.com/m/4cIKs1iwXMS1[agreed] that for issues that list multiple releases in their _Fix Version/s_ field, CLOSE the issue on the release of any of the versions listed; subsequent change to the issue must happen in a new JIRA.
We agreed that for issues that list multiple releases in their _Fix Version/s_ field, CLOSE the issue on the release of any of the versions listed; subsequent change to the issue must happen in a new JIRA.
[[no.permanent.state.in.zk]]
.Only transient state in ZooKeeper!
Expand Down Expand Up @@ -103,7 +103,7 @@ Owners do not need to be committers.
[[hbase.commit.msg.format]]
== Commit Message format
We link:http://search-hadoop.com/m/Gwxwl10cFHa1[agreed] to the following SVN commit message format:
We agreed to the following Git commit message format:
[source]
----
HBASE-xxxxx <title>. (<contributor>)
Expand Down
4 changes: 1 addition & 3 deletions src/main/asciidoc/_chapters/compression.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,6 @@
NOTE: Codecs mentioned in this section are for encoding and decoding data blocks or row keys.
For information about replication codecs, see <<cluster.replication.preserving.tags,cluster.replication.preserving.tags>>.
Some of the information in this section is pulled from a link:http://search-hadoop.com/m/lL12B1PFVhp1/v=threaded[discussion] on the HBase Development mailing list.
HBase supports several different compression algorithms which can be enabled on a ColumnFamily.
Data block encoding attempts to limit duplication of information in keys, taking advantage of some of the fundamental designs and patterns of HBase, such as sorted row keys and the schema of a given table.
Compressors reduce the size of large, opaque byte arrays in cells, and can significantly reduce the storage space needed to store uncompressed data.
Expand Down Expand Up @@ -129,7 +127,7 @@ Come and write the dev list if you are interesting in carrying-on this encoding.
The compression or codec type to use depends on the characteristics of your data. Choosing the wrong type could cause your data to take more space rather than less, and can have performance implications.
In general, you need to weigh your options between smaller size and faster compression/decompression. Following are some general guidelines, expanded from a discussion at link:http://search-hadoop.com/m/lL12B1PFVhp1[Documenting Guidance on compression and codecs].
In general, you need to weigh your options between smaller size and faster compression/decompression. Following are some general guidelines, expanded from a discussion at link:https://lists.apache.org/thread.html/481e67a61163efaaf4345510447a9244871a8d428244868345a155ff%401378926618%40%3Cdev.hbase.apache.org%3E[Documenting Guidance on compression and codecs].
* If you have long keys (compared to the values) or many columns, use a prefix encoder.
FAST_DIFF is recommended, as more testing is needed for Prefix Tree encoding.
Expand Down
15 changes: 9 additions & 6 deletions src/main/asciidoc/_chapters/configuration.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -99,12 +99,12 @@ This section lists required services and some required system configuration.
|JDK 8
|1.1
|link:http://search-hadoop.com/m/DHED4Zlz0R1[Not Supported]
|Not Supported
|yes
|Running with JDK 8 will work but is not well tested.
|1.0
|link:http://search-hadoop.com/m/DHED4Zlz0R1[Not Supported]
|Not Supported
|yes
|Running with JDK 8 will work but is not well tested.
Expand Down Expand Up @@ -306,8 +306,7 @@ HBase-0.94 can additionally work with Hadoop-0.23.x and 2.x, but you may have to
As of Apache HBase 0.96.x, Apache Hadoop 1.0.x at least is required.
Hadoop 2 is strongly encouraged (faster but also has fixes that help MTTR). We will no longer run properly on older Hadoops such as 0.20.205 or branch-0.20-append.
Do not move to Apache HBase 0.96.x if you cannot upgrade your Hadoop. See link:http://search-hadoop.com/m/7vFVx4EsUb2[HBase, mail # dev - DISCUSS:
Have hbase require at least hadoop 1.0.0 in hbase 0.96.0?]
Do not move to Apache HBase 0.96.x if you cannot upgrade your Hadoop. See link:https://lists.apache.org/thread.html/5d5cdf7ccf7aea88398971a221ef3fc65d2b28712dfdfdd649079a35%401330716274%40%3Cdev.hbase.apache.org%3E[HBase, mail # dev - DISCUSS: Have hbase require at least hadoop 1.0.0 in hbase 0.96.0?]
[[hadoop.older.versions]]
==== Hadoop versions 0.20.x - 1.x
Expand Down Expand Up @@ -909,8 +908,12 @@ If your working set it such that block cache does you no good, at least size the
==== link:http://en.wikipedia.org/wiki/Nagle's_algorithm[Nagle's] or the small package problem
If a big 40ms or so occasional delay is seen in operations against HBase, try the Nagles' setting.
For example, see the user mailing list thread, link:http://search-hadoop.com/m/pduLg2fydtE/Inconsistent+scan+performance+with+caching+set+&subj=Re+Inconsistent+scan+performance+with+caching+set+to+1[Inconsistent scan performance with caching set to 1] and the issue cited therein where setting `notcpdelay` improved scan speeds.
You might also see the graphs on the tail of link:https://issues.apache.org/jira/browse/HBASE-7008[HBASE-7008 Set scanner caching to a better default] where our Lars Hofhansl tries various data sizes w/ Nagle's on and off measuring the effect.
For example, see the user mailing list thread,
link:https://lists.apache.org/thread.html/3d7ceb41c04a955b1b1c80480cdba95208ca3e97bf6895a40e0c1bbb%401346186127%40%3Cuser.hbase.apache.org%3E[Inconsistent scan performance with caching set to 1]
and the issue cited therein where setting `notcpdelay` improved scan speeds. You might also see the
graphs on the tail of
link:https://issues.apache.org/jira/browse/HBASE-7008[HBASE-7008 Set scanner caching to a better default]
where our Lars Hofhansl tries various data sizes w/ Nagle's on and off measuring the effect.
[[mttr]]
==== Better Mean Time to Recover (MTTR)
Expand Down
14 changes: 6 additions & 8 deletions src/main/asciidoc/_chapters/developer.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ Being familiar with these guidelines will help the HBase committers to use your
Apache HBase gets better only when people contribute! If you are looking to contribute to Apache HBase, look for link:https://issues.apache.org/jira/issues/?jql=project%20%3D%20HBASE%20AND%20labels%20in%20(beginner)%20AND%20status%20in%20(Open%2C%20%22In%20Progress%22%2C%20Reopened)[issues in JIRA tagged with the label 'beginner'].
These are issues HBase contributors have deemed worthy but not of immediate priority and a good way to ramp on HBase internals.
See link:http://search-hadoop.com/m/DHED43re96[What label
See link:https://lists.apache.org/thread.html/b122265f4e4054cf08f8cd38609fb06af72f398c44f9086b05ef4e21%401407246237%40%3Cdev.hbase.apache.org%3E[What label
is used for issues that are good on ramps for new contributors?] from the dev mailing list for background.
Before you get started submitting code to HBase, please refer to <<developing,developing>>.
Expand Down Expand Up @@ -861,7 +861,7 @@ requirements of the ASF policy on releases.
____

Regards the latter, run `mvn apache-rat:check` to verify all files are suitably licensed.
See link:http://search-hadoop.com/m/DHED4dhFaU[HBase, mail # dev - On recent discussion clarifying ASF release policy]
See link:https://mail-archives.apache.org/mod_mbox/hbase-dev/201406.mbox/%3CCA%2BRK%3D_B8EP0JMFV%2Bdt-k1g%3DBmedzyq2z1GSqrnMMiH6%3DcdoiAA%40mail.gmail.com%3E[HBase, mail # dev - On recent discussion clarifying ASF release policy]
for how we arrived at this process.

[[documentation]]
Expand Down Expand Up @@ -2139,10 +2139,8 @@ We've established the practice of committing to master and then cherry picking b

When there is a minor conflict we can fix it up and just proceed with the commit.
The resulting commit retains the original author.
When the amending author is different from the original committer, add notice of this at the end of the commit message as: `Amending-Author: Author
<committer&apache>` See discussion at link:http://search-hadoop.com/m/DHED4wHGYS[HBase, mail # dev
- [DISCUSSION] Best practice when amending commits cherry picked
from master to branch].
When the amending author is different from the original committer, add notice of this at the end of the commit message as:
`Amending-Author: Author <committer&apache>` See discussion at link:https://lists.apache.org/thread.html/7f514da8b24785a4a64b9f023692ce1b345eba215e1f622820551ca1%401401821274%40%3Cdev.hbase.apache.org%3E[HBase, mail # dev - [DISCUSSION] Best practice when amending commits cherry picked from master to branch].

[[committer.tests]]
====== Committers are responsible for making sure commits do not break the build or tests
Expand All @@ -2154,7 +2152,7 @@ A committer should.
[[git.patch.flow]]
====== Patching Etiquette

In the thread link:http://search-hadoop.com/m/DHED4EiwOz[HBase, mail # dev - ANNOUNCEMENT: Git Migration In Progress (WAS =>
In the thread link:https://lists.apache.org/thread.html/186fcd5eb71973a7b282ecdba41606d3d221efd505d533bb729e1fad%401400648690%40%3Cdev.hbase.apache.org%3E[HBase, mail # dev - ANNOUNCEMENT: Git Migration In Progress (WAS =>
Re: Git Migration)], it was agreed on the following patch flow

. Develop and commit the patch against master first.
Expand All @@ -2176,7 +2174,7 @@ However any substantive discussion (as with any off-list project-related discuss

==== Do not edit JIRA comments

Misspellings and/or bad grammar is preferable to the disruption a JIRA comment edit causes: See the discussion at link:http://search-hadoop.com/?q=%5BReopened%5D+%28HBASE-451%29+Remove+HTableDescriptor+from+HRegionInfo&fc_project=HBase[Re:(HBASE-451) Remove HTableDescriptor from HRegionInfo]
Misspellings and/or bad grammar is preferable to the disruption a JIRA comment edit causes.

[[hbase.archetypes.development]]
=== Development of HBase-related Maven archetypes
Expand Down
1 change: 0 additions & 1 deletion src/main/asciidoc/_chapters/ops_mgt.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -1971,7 +1971,6 @@ Physical data size on disk is distinct from logical size of your data and is aff
See <<regions.arch,regions.arch>>.

* Decreased by <<compression,compression>> and data block encoding, depending on data.
See also link:http://search-hadoop.com/m/lL12B1PFVhp1[this thread].
You might want to test what compression and encoding (if any) make sense for your data.
* Increased by size of region server <<wal,wal>> (usually fixed and negligible - less than half of RS memory size, per RS).
* Increased by HDFS replication - usually x3.
Expand Down
3 changes: 1 addition & 2 deletions src/main/asciidoc/_chapters/performance.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -572,7 +572,6 @@ See <<precreate.regions>>, as well as <<perf.configurations>>
== Reading from HBase
The mailing list can help if you are having performance issues.
For example, here is a good general thread on what to look at addressing read-time issues: link:http://search-hadoop.com/m/qOo2yyHtCC1[HBase Random Read latency > 100ms]
[[perf.hbase.client.caching]]
=== Scan Caching
Expand Down Expand Up @@ -802,7 +801,7 @@ See the link:https://issues.apache.org/jira/browse/HDFS-1599[Umbrella Jira Ticke
Since Hadoop 1.0.0 (also 0.22.1, 0.23.1, CDH3u3 and HDP 1.0) via link:https://issues.apache.org/jira/browse/HDFS-2246[HDFS-2246], it is possible for the DFSClient to take a "short circuit" and read directly from the disk instead of going through the DataNode when the data is local.
What this means for HBase is that the RegionServers can read directly off their machine's disks instead of having to open a socket to talk to the DataNode, the former being generally much faster.
See JD's link:http://files.meetup.com/1350427/hug_ebay_jdcryans.pdf[Performance Talk].
Also see link:http://search-hadoop.com/m/zV6dKrLCVh1[HBase, mail # dev - read short circuit] thread for more discussion around short circuit reads.
Also see link:https://lists.apache.org/thread.html/ce2ce3a3bbd20806d0c017b2e7528e78a46ccb87c063831db051949d%401347548325%40%3Cdev.hbase.apache.org%3E[HBase, mail # dev - read short circuit] thread for more discussion around short circuit reads.
To enable "short circuit" reads, it will depend on your version of Hadoop.
The original shortcircuit read patch was much improved upon in Hadoop 2 in link:https://issues.apache.org/jira/browse/HDFS-347[HDFS-347].
Expand Down
4 changes: 2 additions & 2 deletions src/main/asciidoc/_chapters/schema_design.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -190,7 +190,7 @@ If your rows and column names are large, especially compared to the size of the
One such is the case described by Marc Limotte at the tail of link:https://issues.apache.org/jira/browse/HBASE-3551?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13005272#comment-13005272[HBASE-3551] (recommended!). Therein, the indices that are kept on HBase storefiles (<<hfile>>) to facilitate random access may end up occupying large chunks of the HBase allotted RAM because the cell value coordinates are large.
Mark in the above cited comment suggests upping the block size so entries in the store file index happen at a larger interval or modify the table schema so it makes for smaller rows and column names.
Compression will also make for larger indices.
See the thread link:http://search-hadoop.com/m/hemBv1LiN4Q1/a+question+storefileIndexSize&subj=a+question+storefileIndexSize[a question storefileIndexSize] up on the user mailing list.
See the thread link:https://lists.apache.org/thread.html/b158eae5d8888d3530be378298bca90c17f80982fdcdfa01d0844c3d%401306240189%40%3Cuser.hbase.apache.org%3E[a question storefileIndexSize] up on the user mailing list.
Most of the time small inefficiencies don't matter all that much. Unfortunately, this is a case where they do.
Whatever patterns are selected for ColumnFamilies, attributes, and rowkeys they could be repeated several billion times in your data.
Expand Down Expand Up @@ -581,7 +581,7 @@ However, HBase scales better at larger data volumes, so this is a feature trade-
Pay attention to <<performance>> when implementing any of these approaches.
Additionally, see the David Butler response in this dist-list thread link:http://search-hadoop.com/m/nvbiBp2TDP/Stargate%252Bhbase&subj=Stargate+hbase[HBase, mail # user - Stargate+hbase]
Additionally, see the David Butler response in this dist-list thread link:https://lists.apache.org/thread.html/b0ca33407f010d5b1be67a20d1708e8d8bb1e147770f2cb7182a2e37%401300972712%40%3Cuser.hbase.apache.org%3E[HBase, mail # user - Stargate+hbase]
[[secondary.indexes.filter]]
=== Filter Query
Expand Down
Loading

0 comments on commit e1c9299

Please sign in to comment.