Uploaded image for project: 'CDAP'
  1. CDAP
  2. CDAP-13575

not able to test the logistic Regression model or even decision tree model.

    Details

    • Type: Bug
    • Status: Resolved
    • Priority: Blocker
    • Resolution: Duplicate
    • Affects Version/s: 4.3.4
    • Fix Version/s: None
    • Component/s: CDAP Examples, Market
    • Labels:
      None
    • Rank:
      1|i00dkn:

      Description

      2018-06-19 12:02:39,695 - ERROR [Executor task launch worker-0:o.a.s.e.Executor@95] - Exception in task 1.0 in stage 2.0 (TID 4)
      org.apache.parquet.hadoop.BadConfigurationException: class org.apache.spark.sql.execution.datasources.parquet.CatalystReadSupport set in job conf at parquet.read.support.class is not a subclass of org.apache.parquet.hadoop.api.ReadSupport
      at org.apache.parquet.hadoop.util.ConfigurationUtil.getClassFromConfig(ConfigurationUtil.java:35) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.parquet.hadoop.ParquetInputFormat.getReadSupportClass(ParquetInputFormat.java:182) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.parquet.hadoop.ParquetInputFormat.getReadSupport(ParquetInputFormat.java:257) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.parquet.hadoop.ParquetInputFormat.createRecordReader(ParquetInputFormat.java:245) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.spark.rdd.SqlNewHadoopRDD$$anon$1.<init>(SqlNewHadoopRDD.scala:178) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.SqlNewHadoopRDD.compute(SqlNewHadoopRDD.scala:126) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.scheduler.Task.run(Task.scala:89) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_121]
      at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_121]
      at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
      2018-06-19 12:02:39,703 - ERROR [Executor task launch worker-1:o.a.s.e.Executor@95] - Exception in task 0.0 in stage 2.0 (TID 3)
      org.apache.parquet.hadoop.BadConfigurationException: class org.apache.spark.sql.execution.datasources.parquet.CatalystReadSupport set in job conf at parquet.read.support.class is not a subclass of org.apache.parquet.hadoop.api.ReadSupport
      at org.apache.parquet.hadoop.util.ConfigurationUtil.getClassFromConfig(ConfigurationUtil.java:35) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.parquet.hadoop.ParquetInputFormat.getReadSupportClass(ParquetInputFormat.java:182) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.parquet.hadoop.ParquetInputFormat.getReadSupport(ParquetInputFormat.java:257) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.parquet.hadoop.ParquetInputFormat.createRecordReader(ParquetInputFormat.java:245) ~[parquet-hadoop-1.8.1.jar:1.8.1]
      at org.apache.spark.rdd.SqlNewHadoopRDD$$anon$1.<init>(SqlNewHadoopRDD.scala:178) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.SqlNewHadoopRDD.compute(SqlNewHadoopRDD.scala:126) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.scheduler.Task.run(Task.scala:89) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) ~[co.cask.cdap.spark-assembly-1.6.1.jar:na]
      at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_121]
      at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_121]
      at java.lang.Thread.run(Thread.java:745) [na:1.8.0_121]
      2018-06-19 12:02:39,794 - WARN [task-result-getter-0:o.a.s.s.TaskSetManager@70] - Lost task 0.0 in stage 2.0 (TID 3, localhost): org.apache.parquet.hadoop.BadConfigurationException: class org.apache.spark.sql.execution.datasources.parquet.CatalystReadSupport set in job conf at parquet.read.support.class is not a subclass of org.apache.parquet.hadoop.api.ReadSupport
      at org.apache.parquet.hadoop.util.ConfigurationUtil.getClassFromConfig(ConfigurationUtil.java:35)
      at org.apache.parquet.hadoop.ParquetInputFormat.getReadSupportClass(ParquetInputFormat.java:182)
      at org.apache.parquet.hadoop.ParquetInputFormat.getReadSupport(ParquetInputFormat.java:257)
      at org.apache.parquet.hadoop.ParquetInputFormat.createRecordReader(ParquetInputFormat.java:245)
      at org.apache.spark.rdd.SqlNewHadoopRDD$$anon$1.<init>(SqlNewHadoopRDD.scala:178)
      at org.apache.spark.rdd.SqlNewHadoopRDD.compute(SqlNewHadoopRDD.scala:126)
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
      at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
      at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
      at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
      at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
      at org.apache.spark.scheduler.Task.run(Task.scala:89)
      at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
      at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
      at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
      at java.lang.Thread.run(Thread.java:745)

        Attachments

          Issue Links

            Activity

              People

              • Assignee:
                bhooshan Bhooshan Mogal
                Reporter:
                vivek.angadi vivek
              • Votes:
                0 Vote for this issue
                Watchers:
                2 Start watching this issue

                Dates

                • Created:
                  Updated:
                  Resolved: