Register / Login
Win a copy of
The Java Performance Companion
this week in the
Hive Gzip Compression splitting supported now?
posted 3 years ago
Does Hadoop automatically support splitting Gzip files into Blocks now? I have read that splitting doesn't work for tables using gzip compression in Hadoop/Hive here:
From the above link: "in this case Hadoop will not be able to split your file into chunks/blocks and run multiple maps in parallel. This can cause under-utilization of your cluster's 'mapping' power."
However, when I load my table exactly as they describe, I notice that the gz I load is definately split up into blocks in the place it stores my HDFS files. It looks like this after doing the load:
[current]$ pwd /foo/dev/hadoop/HDFS/dfs/data/current [current]$ ls -la -rw-r--r--. 1 foo bar 671 Dec 19 15:34 blk_105489922789526087_1136.meta -rw-r--r--. 1 foo bar 67108864 Dec 19 14:26 blk_-1527019105370167199 -rw-r--r--. 1 foo bar 524295 Dec 19 14:26 blk_-1527019105370167199_1075.meta -rw-r--r--. 1 foo bar 67108864 Dec 19 14:26 blk_-226975864542913836 -rw-r--r--. 1 foo bar 524295 Dec 19 14:26 blk_-226975864542913836_1075.meta -rw-r--r--. 1 foo bar 67108864 Dec 19 14:26 blk_-2476094208541673790
It is clearly chopping it up into 64 mb blocks during the load to HDFS.
Is this something they have added recently? I'm using Hadoop 1.0.4, r1393290 in psuedo cluster mode.
text file extractor
Issues with Tortoise CVS and Cruise control
How to Enable SSL on Tomcat 7 on Linux?
java.io.FileNotFoundException: Too many open files