Cassandra和HBase都不能有效地支持像图像一样的blob存储.直接存储在HDFS上会强调Namenode.Facebook的Haystack不是开源的.那么Lustre是分布式blob存储的不错选择吗?还有什么建议?
我已经安装了hadoop和hbase cdh3u2.在hadoop我有一个文件在路径上/home/file.txt.它有像这样的数据
one,1
two,2
three,3
Run Code Online (Sandbox Code Playgroud)
我想将此文件导入hbase.在那里,第一个字段应解析为String,第二个字段解析为整数,然后它应推入hbase.帮我这样做
athanks in dvance ....
我想在hbase中清空一个表...例如:user.是否有任何命令或函数来清空表而不删除它...
我的表结构是:
$mutations = array(
new Mutation( array(
'column' => 'username:1',
'value' =>$name
) ),
new Mutation( array(
'column' => 'email:1',
'value' =>$email
) )
);
$hbase->mutateRow("user",$key,$mutations);
Run Code Online (Sandbox Code Playgroud)
有人能帮我吗?
我有点卡住修复故障表(在Hbase 0.92.1-cdh4.0.0,Hadoop 2.0.0-cdh4.0.0上)
转换中的某个区域未完成:
Region State
bf2025f4bc154914b5942af4e72ea063 counter_traces,1329773878.35_766a0b4df75e4381a686fbc07db9e333,1339425291230.bf2025f4bc154914b5942af4e72ea063. state=OFFLINE, ts=Tue Jun 12 11:43:53 CEST 2012 (0s ago), server=null
Run Code Online (Sandbox Code Playgroud)
当我跑步时sudo -u hbase hbase hbck -repair,我明白了:
Number of empty REGIONINFO_QUALIFIER rows in .META.: 0
ERROR: Region { meta => counter_traces,1329773878.35_766a0b4df75e4381a686fbc07db9e333,1339425291230.bf2025f4bc154914b5942af4e72ea063., hdfs => hdfs://hbase001:8020/hbase/counter_traces/bf2025f4bc154914b5942af4e72ea063, deployed => } not deployed on any region server.
Trying to fix unassigned region...
12/06/12 11:44:40 INFO util.HBaseFsckRepair: Region still in transition, waiting for it to become assigned: {NAME => 'counter_traces,1329773878.35_766a0b4df75e4381a686fbc07db9e333,1339425291230.bf2025f4bc154914b5942af4e72ea063.', STARTKEY => '1329773878.35_766a0b4df75e4381a686fbc07db9e333', ENDKEY => '1329793347.58_163865765c0a11e184ab003048f0e77e', …Run Code Online (Sandbox Code Playgroud) 我在下载后使用ubuntu 12.04并取消了tar-hbase-0.94.1.jar.我通过以下命令在独立模式下启动了hbash
user@ubuntu:~/temp/hbase-0.94.1$ bin/start-hbase.sh
starting master, logging to /home/user/temp/hbase-0.94.1/bin/../logs/hbase-user-master-ubuntu.out
Run Code Online (Sandbox Code Playgroud)
但是当我尝试访问hbash状态UI时
http://localhost:60010
Run Code Online (Sandbox Code Playgroud)
,我得到"无法连接".
谢谢.
担
有没有办法在扫描期间指定多个过滤器?例如 - 指定a ColumnFamilyFilter和RowFilter?
Filter rowFilter =
new RowFilter(CompareFilter.CompareOp.EQUAL, new RegexStringComparator(
rowFilterString));
Scan s = new Scan();
s.setFilter(rowFilter);
Run Code Online (Sandbox Code Playgroud)
我想也是一个补充ColumnFilter到s.但它显然会覆盖最新的过滤器.
我安装了Hadoop2.2.0和Hbase0.98.0,这就是我的工作:
$ ./bin/start-hbase.sh
$ ./bin/hbase shell
2.0.0-p353 :001 > list
Run Code Online (Sandbox Code Playgroud)
然后我得到了这个:
ERROR: Can't get master address from ZooKeeper; znode data == null
Run Code Online (Sandbox Code Playgroud)
为什么我收到此错误?另一个问题:我需要运行./sbin/start-dfs.sh和./sbin/start-yarn.sh之前我跑基地?
还有什么用于./sbin/start-dfs.sh和./sbin/start-yarn.sh?
以下是我的一些conf doc:
<configuration>
<property>
<name>hbase.rootdir</name>
<value>hdfs://127.0.0.1:9000/hbase</value>
</property>
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
</property>
<property>
<name>hbase.tmp.dir</name>
<value>/Users/apple/Documents/tools/hbase-tmpdir/hbase-data</value>
</property>
<property>
<name>hbase.zookeeper.quorum</name>
<value>localhost</value>
</property>
<property>
<name>hbase.zookeeper.property.dataDir</name>
<value>/Users/apple/Documents/tools/hbase-zookeeper/zookeeper</value>
</property>
</configuration>
Run Code Online (Sandbox Code Playgroud)
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost:9000</value>
<description>The name of the default file system.</description>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/Users/micmiu/tmp/hadoop</value>
<description>A base for other temporary directories.</description>
</property> …Run Code Online (Sandbox Code Playgroud) 我在HBase上使用OpenTSDB(虚拟盒上的伪分布式Hadoop)以非常高的负载(~50,000记录/秒)发送数据.系统工作了一段时间,但突然下降了.我终止了OpenTSDB和HBase.不幸的是,我再也不能把它们搞砸了.每次我尝试运行HBase和OpenTSDB时,都会显示错误日志.在这里我列出了日志:
RegionServer的:
2015-07-01 18:15:30,752 INFO [sync.3] wal.FSHLog: Slow sync cost: 112 ms, current pipeline: [192.168.56.101:50010]
2015-07-01 18:15:41,277 INFO [regionserver/node1.vmcluster/192.168.56.101:16201.logRoller] wal.FSHLog: Rolled WAL /hbase/WALs/node1.vmcluster,16201,1435738612093/node1.vmcluster%2C16201%2C1435738612093.default.1435742101122 with entries=3841, filesize=123.61 MB; new WAL /hbase/WALs/node1.vmcluster,16201,1435738612093/node1.vmcluster%2C16201%2C1435738612093.default.1435742141109
2015-07-01 18:15:41,278 INFO [regionserver/node1.vmcluster/192.168.56.101:16201.logRoller] wal.FSHLog: Archiving hdfs://node1.vmcluster:9000/hbase/WALs/node1.vmcluster,16201,1435738612093/node1.vmcluster%2C16201%2C1435738612093.default.1435742061805 to hdfs://node1.vmcluster:9000/hbase/oldWALs/node1.vmcluster%2C16201%2C1435738612093.default.1435742061805
2015-07-01 18:15:42,249 INFO [MemStoreFlusher.0] regionserver.HRegion: Started memstore flush for tsdb,,1435740133573.1a692e2668a2b4a71aaf2805f9b00a72., current region memstore size 132.20 MB
2015-07-01 18:15:42,381 INFO [MemStoreFlusher.1] regionserver.HRegion: Started memstore flush for tsdb,,1435740133573.1a692e2668a2b4a71aaf2805f9b00a72., current region memstore size 133.09 MB
2015-07-01 18:15:42,382 WARN [MemStoreFlusher.1] regionserver.DefaultMemStore: Snapshot called again without …Run Code Online (Sandbox Code Playgroud)