Cassandra的数据分布情况测试

1、规划种子节点ip和Token值的对应

4个种子节点ip
	192.168.0.231
	192.168.0.232
	192.168.0.233
	192.168.0.234

	进入python,计算Token
	#192.168.0.231对应的Token为
	>>> print 2 ** 127 / 4 * 1 
	42535295865117307932921825928971026432

	#192.168.0.232对应的Token为
	>>> print 2 ** 127 / 4 * 2
	85070591730234615865843651857942052864

	#192.168.0.233对应的Token为
	>>> print 2 ** 127 / 4 * 3
	127605887595351923798765477786913079296

	#192.168.0.234对应的Token为
	>>> print 2 ** 127 / 4 * 4
	170141183460469231731687303715884105728

或者

import java.math.BigInteger;

public class InitToken {
	public static void main(String[] args) {
		int nodes = 4;//节点的总数量
		for (int i = 1; i <= nodes; i++) {
			BigInteger hs = new BigInteger("2");
			BigInteger res = hs.pow(127);
			BigInteger div = res.divide(new BigInteger(nodes + ""));
			BigInteger fin = div.multiply(new BigInteger(i + ""));
			System.out.println(fin);
		}

	}

}
 

2、配置cassandra的每个节点

      a、cassandra/bin/cassandra.in.sh配置本机的jdk安装路径

JAVA_HOME=/usr/local/jdk6
 

      b、cassandra/conf/cassandra.yaml

                cluster_name: 'ots'
		commitlog_directory: cassandra/data/commitlog
		saved_caches_directory: cassandra/data/saved_caches
		data_file_directories:
    			- cassandra/data/data
		#配置种子节点ip列表
		- seeds: "192.168.0.231,192.168.0.232,192.168.0.233,192.168.0.234"
		#上面的配置基本保持一致

		#当前节点的ip(这个ip主要是用来节点和节点之间通讯的ip)
		listen_address: 192.168.0.231
		#当前节点的ip(这个ip主要是用来相应客户端操作的ip)
		rpc_address: 192.168.0.231

3、启动每个节点的cassandra

nohup cassandra/bin/cassandra -f &

4、设置每个节点的Token值

    进入每个节点,把启动时默认生成的Token值改变为我们规划的Token值
    或者直接在配置文件cassandra.yaml中指定Token值来规划,就不用下面的动态规划了。
    如下:

          ./bin/nodetool -h 192.168.0.231 -p 7199 move 42535295865117307932921825928971026432
	 ./bin/nodetool -h 192.168.0.232 -p 7199 move 42535295865117307932921825928971026432
	。。。。。。。。。。。。。。。。。。。

 5、初始化数据存储结构

        客户端连接到集群中的某一个节点,初始化数据结构。运行初始化脚本,如下

./bin/cassandra-cli -h 192.168.0.231 -p 9160

 参考json数据结构模型

{
            "key":{
                     "name":"140 bytes",
                     "cardno":"140 bytes",
                     "ticketno":"140 bytes",
                     "traindate":"140 bytes",
                    "startstation":"140 bytes",
                     "endtstation":"140 bytes",
                     "seatinfo":"140 bytes",
                   }
        }
 

        具体脚本:

                create keyspace user_train;
		use user_train;
		create column family users with comparator=UTF8Type 
		and column_metadata=[{column_name:name,validation_class:UTF8Type,index_type:KEYS},
		{column_name:cardno,validation_class:UTF8Type,index_type:KEYS},
		{column_name:ticketno,validation_class:UTF8Type,index_type:KEYS},
		{column_name:traindate,validation_class:UTF8Type},
		{column_name:startstation,validation_class:UTF8Type},
		{column_name:endtstation,validation_class:UTF8Type},
		{column_name:seatinfo,validation_class:UTF8Type}];

		#分发策略,主要是将存放到一个节点的一份数据,分发到另一个节点一份,节点的选取由Cassandra和配置文件决定。
		update keyspace user_train with strategy_options = {datacenter1:2};

5、使用hector客户端api操作cassandra数据库,观察数据分布

插入数据的时候观察每个节点的数据分布是否均衡。
    首先需要登录每个节点,不停的使用如下命令,刷新数据的分布情况。

./bin/nodetool -h 192.168.0.232 -p 7199 ring 
 


    写入数据代码

Cluster cluster = HFactory
                .getOrCreateCluster("ots",
                        "192.168.0.231:9160," +
                        "192.168.0.232:9160," +
                        "192.168.0.233:9160," +
                        "192.168.0.234:9160");

        Keyspace keyspace = HFactory.createKeyspace("user_train", cluster);

        String ktest = "";
        for (int i = 0; i < 120; i++) {
            ktest += "x";
        }
        try {
            Mutator<String> mutator = HFactory.createMutator(keyspace,
                    stringSerializer);
            long startTime = System.currentTimeMillis();
            for (int i = 0; i < 10000*10; i++) {
                mutator.addInsertion("username" + i, "users",
                        HFactory.createStringColumn("name", ktest+"username" + i))
                        .addInsertion(
                                "username" + i,
                                "users",
                                HFactory.createStringColumn("cardno", ktest+"cardno"
                                        + i))
                        .addInsertion(
                                "username" + i,
                                "users",
                                HFactory.createStringColumn("ticketno",
                                        ktest+"ticketno" + i))
                        .addInsertion(
                                "username" + i,
                                "users",
                                HFactory.createStringColumn("traindate",
                                        ktest+"traindate" + i))
                        .addInsertion(
                                "username" + i,
                                "users",
                                HFactory.createStringColumn("startstation",
                                        ktest+"startstation" + i))
                        .addInsertion(
                                "username" + i,
                                "users",
                                HFactory.createStringColumn("endtstation",
                                        ktest+"endtstation" + i))
                        .addInsertion(
                                "username" + i,
                                "users",
                                HFactory.createStringColumn("seatinfo",
                                        ktest+"seatinfo" + i));
                if (i % 500 == 0) {
                    mutator.execute();
                    System.out.println(i);
                }
            }
            mutator.execute();
            System.out.println("insert time: " + (System.currentTimeMillis() - startTime));
        } catch (HectorException he) {
            he.printStackTrace();
        }
        cluster.getConnectionManager().shutdown();
   

猜你喜欢

转载自xiaofancn.iteye.com/blog/1455339