BugsCode 2018-03-02 07:15 采纳率: 0%
浏览 1483
已结题

CarbonData1.3.0导入数据出现锁问题

scala> carbon.sql("CREATE TABLE IF NOT EXISTS test_table(id string,name string,city string,age Int) STORED BY 'carbondata'")
18/03/01 22:35:12 AUDIT CarbonCreateTableCommand: [localhost.localdomain][root][Thread-1]Creating Table with Database name [default] and Table name [test_table]
res9: org.apache.spark.sql.DataFrame = []

scala> carbon.sql("SELECT * FROM test_table").show()
+---+----+----+---+
| id|name|city|age|
+---+----+----+---+
+---+----+----+---+

scala> carbon.sql("LOAD DATA INPATH '/opt/hadoop/sample.csv INTO TABLE test_table")
18/03/01 22:43:51 ERROR CarbonLoaderUtil: main Not able to acquire the lock for Table status updation for table default.test_table
18/03/01 22:43:51 ERROR CarbonLoaderUtil: main Unable to unlock Table lock for tabledefault.test_table during table status updation
18/03/01 22:43:51 ERROR CarbonLoadDataCommand: main
java.io.IOException: Dataload failed due to failure in table status updation for test_table
at org.apache.carbondata.processing.util.CarbonLoaderUtil.readAndUpdateLoadProgressInTableMeta(CarbonLoaderUtil.java:443)
at org.apache.carbondata.processing.util.CarbonLoaderUtil.readAndUpdateLoadProgressInTableMeta(CarbonLoaderUtil.java:449)
at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:236)
at org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:92)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:67)
at org.apache.spark.sql.Dataset.(Dataset.scala:183)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:68)
at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:632)
at $line40.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.(:33)
at $line40.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.(:38)
at $line40.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.(:40)
at $line40.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw.(:42)
at $line40.$read$$iw$$iw$$iw$$iw$$iw$$iw.(:44)
at $line40.$read$$iw$$iw$$iw$$iw$$iw.(:46)
at $line40.$read$$iw$$iw$$iw$$iw.(:48)
at $line40.$read$$iw$$iw$$iw.(:50)
at $line40.$read$$iw$$iw.(:52)
at $line40.$read$$iw.(:54)
at $line40.$read.(:56)
at $line40.$read$.(:60)
at $line40.$read$.()
at $line40.$eval$.$print$lzycompute(:7)
at $line40.$eval$.$print(:6)
at $line40.$eval.$print()
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786)
at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047)
at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:638)
at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:637)
at scala.reflect.internal.util.ScalaClassLoader$class.asContext(ScalaClassLoader.scala:31)
at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:19)
at scala.tools.nsc.interpreter.IMain$WrappedRequest.loadAndRunReq(IMain.scala:637)
at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:569)
at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565)
at scala.tools.nsc.interpreter.ILoop.interpretStartingWith(ILoop.scala:807)
at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681)
at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395)
at scala.tools.nsc.interpreter.ILoop.loop(ILoop.scala:415)
at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply$mcZ$sp(ILoop.scala:923)
at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
at scala.reflect.internal.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:97)
at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909)
at org.apache.spark.repl.Main$.doMain(Main.scala:74)
at org.apache.spark.repl.Main$.main(Main.scala:54)
at org.apache.spark.repl.Main.main(Main.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:775)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
18/03/01 22:43:51 AUDIT CarbonLoadDataCommand: [localhost.localdomain][root][Thread-1]Dataload failure for default.test_table. Please check the logs
18/03/01 22:43:51 ERROR CarbonLoadDataCommand: main Got exception java.io.IOException: Dataload failed due to failure in table status updation for test_table when processing data. But this command does not support undo yet, skipping the undo part.
java.io.IOException: Dataload failed due to failure in table status updation for test_table
at org.apache.carbondata.processing.util.CarbonLoaderUtil.readAndUpdateLoadProgressInTableMeta(CarbonLoaderUtil.java:443)
at org.apache.carbondata.processing.util.CarbonLoaderUtil.readAndUpdateLoadProgressInTableMeta(CarbonLoaderUtil.java:449)
at org.apache.spark.sql.execution.command.management.CarbonLoadDataCommand.processData(CarbonLoadDataCommand.scala:236)
at org.apache.spark.sql.execution.command.AtomicRunnableCommand.run(package.scala:92)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:67)
at org.apache.spark.sql.Dataset.(Dataset.scala:183)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:68)
at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:632)
... 50 elided

  • 写回答

1条回答 默认 最新

  • BugsCode 2018-03-02 08:52
    关注

    官方解决锁方案:

    import org.apache.carbondata.core.util.CarbonProperties
    import org.apache.carbondata.core.constants.CarbonCommonConstants
    CarbonProperties.getInstance().addProperty(CarbonCommonConstants.LOCK_TYPE, "HDFSLOCK")

    再把自己的读取方式由
    carbon.sql("LOAD DATA INPATH '/opt/hadoop/sample.csv' INTO TABLE test_table")
    改为
    carbon.sql("LOAD DATA INPATH 'hdfs://localhost:9000/user/root/input/sample.csv' INTO TABLE test_table")

    scala> carbon.sql("LOAD DATA INPATH 'hdfs://localhost:9000/user/root/input/sample.csv' INTO TABLE test_table")
    18/03/02 00:46:14 AUDIT CarbonDataRDDFactory$: [localhost.localdomain][root][Thread-1]Data load request has been received for table default.test_table
    18/03/02 00:46:14 WARN CarbonDataProcessorUtil: main sort scope is set to LOCAL_SORT
    18/03/02 00:46:17 WARN CarbonDataProcessorUtil: [Executor task launch worker for task 0][partitionID:table;queryID:1114726938817] sort scope is set to LOCAL_SORT
    18/03/02 00:46:17 WARN CarbonDataProcessorUtil: [Executor task launch worker for task 0][partitionID:table;queryID:1114726938817] batch sort size is set to 0
    18/03/02 00:46:17 WARN CarbonDataProcessorUtil: [Executor task launch worker for task 0][partitionID:table;queryID:1114726938817] sort scope is set to LOCAL_SORT
    18/03/02 00:46:17 WARN CarbonDataProcessorUtil: [Executor task launch worker for task 0][partitionID:table;queryID:1114726938817] Error occurs while creating dirs: /tmp/carbon1114860648927_0/Fact/Part0/Segment_0/0
    18/03/02 00:46:17 WARN CarbonDataProcessorUtil: [Executor task launch worker for task 0][partitionID:table;queryID:1114726938817] sort scope is set to LOCAL_SORT
    18/03/02 00:46:20 AUDIT CarbonDataRDDFactory$: [localhost.localdomain][root][Thread-1]Data load is successful for default.test_table
    res3: org.apache.spark.sql.DataFrame = []

    scala> carbon.sql("SELECT * FROM test_table").show()
    +---+----+----+---+
    | id|name|city|age|
    +---+----+----+---+
    | 1|aaaa| xm| 20|
    | 2|bbbb| xm| 21|
    | 3|cccc| zz| 30|
    | 4|dddd| hh| 20|
    +---+----+----+---+

    评论

报告相同问题?

悬赏问题

  • ¥15 如何在scanpy上做差异基因和通路富集?
  • ¥20 关于#硬件工程#的问题,请各位专家解答!
  • ¥15 关于#matlab#的问题:期望的系统闭环传递函数为G(s)=wn^2/s^2+2¢wn+wn^2阻尼系数¢=0.707,使系统具有较小的超调量
  • ¥15 FLUENT如何实现在堆积颗粒的上表面加载高斯热源
  • ¥30 截图中的mathematics程序转换成matlab
  • ¥15 动力学代码报错,维度不匹配
  • ¥15 Power query添加列问题
  • ¥50 Kubernetes&Fission&Eleasticsearch
  • ¥15 報錯:Person is not mapped,如何解決?
  • ¥15 c++头文件不能识别CDialog