1

我正在尝试向 cassandra 集群插入一个包含几列大小为 500MB 的单行,但我遇到了错误。

com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (tried: [/10.138.90.207:9042, /10.138.90.208:9042, /10.138.90.191:9042, /10.138.90.240:9042, /10.138.90.232:9042, /10.138.90.205:9042, /10.138.90.236:9042, /10.138.90.246:9042] - use getErrors() for details)
at com.datastax.driver.core.exceptions.NoHostAvailableException.copy(NoHostAvailableException.java:65)
at com.datastax.driver.core.DefaultResultSetFuture.extractCauseFromExecutionException(DefaultResultSetFuture.java:256)
at com.datastax.driver.core.DefaultResultSetFuture.getUninterruptibly(DefaultResultSetFuture.java:172)
at com.datastax.driver.core.AbstractSession.execute(AbstractSession.java:52)
at com.tcs.asml.cassandra.Crud.Insert(Crud.java:44)
at com.tcs.asml.factory.PartToolInsert.main(PartToolInsert.java:68)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.eclipse.jdt.internal.jarinjarloader.JarRsrcLoader.main(JarRsrcLoader.java:58)
Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (tried: [/10.138.90.207:9042, /10.138.90.208:9042, /10.138.90.191:9042, /10.138.90.240:9042, /10.138.90.232:9042, /10.138.90.205:9042, /10.138.90.236:9042, /10.138.90.246:9042] - use getErrors() for details)
at com.datastax.driver.core.RequestHandler.sendRequest(RequestHandler.java:103)
at com.datastax.driver.core.RequestHandler$1.run(RequestHandler.java:175)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)

当我在异常中打印获取错误时,它显示集群中所有节点的读取错误期间超时。

Get errors: 
{/10.138.90.207:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read, /10.138.90.191:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read, /10.138.90.208:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read, /10.138.90.240:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read, /10.138.90.232:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read, /10.138.90.205:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read, /10.138.90.236:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read, /10.138.90.246:9042=com.datastax.driver.core.exceptions.DriverException: Timeout during read}

集群详情:

  • 一个数据中心,有 8 个节点,每个节点 16GB RAM
  • 每个节点中的单个硬盘。
  • 所有节点都以 10mbps 带宽和默认延迟连接。

我尝试使用以下命令增加读取时间。

cluster.getConfiguration().getSocketOptions().setReadTimeoutMillis(60000);

下面是现在使用的 yaml 配置。

  • 内存总空间:4Gb
  • 提交日志段大小:512MB
  • read_request_timeout_in_ms (ms): 10000
  • request_timeout_in_ms(毫秒):10000
  • 并发读取:32
  • 并发写入:32

我在尝试插入 250mb 行时遇到了同样的问题,通过将读取时间设置为 30 秒解决了这个问题。

cluster.getConfiguration().getSocketOptions().setReadTimeoutMillis(30000);

但是对于 500MB 的行大小,它不起作用。

谁能给我一些想法如何调整 cassandra 以插入包含大量数据的单行。

谢谢。

4

1 回答 1

0

问:为什么cassandra需要连续存储500MB或200MB的数据?cassandra 中分区大小的最佳选择是 100MB,可能是几百。Cassandra 是一个用于快速存储和快速查询的数据存储。500MB 的数据也不会给你。那么为什么要使用 cassandra 呢?

于 2014-08-21T13:47:39.223 回答