Spark sql and hive causing “Error: Could not find or load main class”












0















I am new to big data. I have developed an spark application using Scala to communicate with hive. It is working fine on Intellij idea. But when I am building an jar file with all dependency I am getting the error




Error: Could not find or load main class




After further debugging I have found that when I am including spark SQL or hive then I am getting the error. It is a maven project. Please have a look on my pom.xml



<modelVersion>4.0.0</modelVersion>

<groupId>spk</groupId>
<artifactId>spk</artifactId>
<version>1.0-SNAPSHOT</version>

<dependencies>
<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-core -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-sql -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!--<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive_2.11</artifactId>
<version>2.0.2</version>
</dependency>-->

<!--<dependency>
<groupId>org.apache.hive</groupId>
<artifactId>hive-exec</artifactId>
<version>1.2.2</version>
</dependency>-->

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka_2.11</artifactId>
<version>2.0.1</version>
</dependency>

<!-- https://mvnrepository.com/artifact/com.101tec/zkclient -->
<dependency>
<groupId>com.101tec</groupId>
<artifactId>zkclient</artifactId>
<version>0.11</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka-clients</artifactId>
<version>0.10.0.0</version>
</dependency>

</dependencies>

<build>
<sourceDirectory>src/main/scala</sourceDirectory>
<plugins>
<plugin>
<groupId>org.scala-tools</groupId>
<artifactId>maven-scala-plugin</artifactId>
<version>2.11</version>
<executions>
<execution>
<goals>
<goal>compile</goal>
<goal>testCompile</goal>
</goals>
<configuration>
<args>
<!--<arg>-make:transitive</arg>-->
<arg>-dependencyfile</arg>
<arg>${project.build.directory}/.scala_dependencies</arg>
</args>
</configuration>
</execution>
</executions>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-surefire-plugin</artifactId>
<version>2.11</version>
<configuration>
<useFile>false</useFile>
<disableXmlReport>true</disableXmlReport>
<!-- If you have classpath issue like NoDefClassError,... -->
<!-- useManifestOnlyJar>false</useManifestOnlyJar -->
<includes>
<include>**/*Test.*</include>
<include>**/*Suite.*</include>
</includes>
</configuration>
</plugin>
</plugins>
</build>


Thanks in advance.










share|improve this question


















  • 1





    What is the command you are using to run your jar?

    – anuj saxena
    Nov 25 '18 at 4:02











  • java -jar jarfile.jar. Building excluding spark SQL is working fine

    – Binod Sarkar
    Nov 25 '18 at 4:12








  • 1





    Use spark-submit to run your jar. Follow the link: spark.apache.org/docs/latest/submitting-applications.html

    – anuj saxena
    Nov 25 '18 at 4:17











  • I have tried the spark-submit also but same error

    – Binod Sarkar
    Nov 25 '18 at 4:21
















0















I am new to big data. I have developed an spark application using Scala to communicate with hive. It is working fine on Intellij idea. But when I am building an jar file with all dependency I am getting the error




Error: Could not find or load main class




After further debugging I have found that when I am including spark SQL or hive then I am getting the error. It is a maven project. Please have a look on my pom.xml



<modelVersion>4.0.0</modelVersion>

<groupId>spk</groupId>
<artifactId>spk</artifactId>
<version>1.0-SNAPSHOT</version>

<dependencies>
<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-core -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-sql -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!--<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive_2.11</artifactId>
<version>2.0.2</version>
</dependency>-->

<!--<dependency>
<groupId>org.apache.hive</groupId>
<artifactId>hive-exec</artifactId>
<version>1.2.2</version>
</dependency>-->

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka_2.11</artifactId>
<version>2.0.1</version>
</dependency>

<!-- https://mvnrepository.com/artifact/com.101tec/zkclient -->
<dependency>
<groupId>com.101tec</groupId>
<artifactId>zkclient</artifactId>
<version>0.11</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka-clients</artifactId>
<version>0.10.0.0</version>
</dependency>

</dependencies>

<build>
<sourceDirectory>src/main/scala</sourceDirectory>
<plugins>
<plugin>
<groupId>org.scala-tools</groupId>
<artifactId>maven-scala-plugin</artifactId>
<version>2.11</version>
<executions>
<execution>
<goals>
<goal>compile</goal>
<goal>testCompile</goal>
</goals>
<configuration>
<args>
<!--<arg>-make:transitive</arg>-->
<arg>-dependencyfile</arg>
<arg>${project.build.directory}/.scala_dependencies</arg>
</args>
</configuration>
</execution>
</executions>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-surefire-plugin</artifactId>
<version>2.11</version>
<configuration>
<useFile>false</useFile>
<disableXmlReport>true</disableXmlReport>
<!-- If you have classpath issue like NoDefClassError,... -->
<!-- useManifestOnlyJar>false</useManifestOnlyJar -->
<includes>
<include>**/*Test.*</include>
<include>**/*Suite.*</include>
</includes>
</configuration>
</plugin>
</plugins>
</build>


Thanks in advance.










share|improve this question


















  • 1





    What is the command you are using to run your jar?

    – anuj saxena
    Nov 25 '18 at 4:02











  • java -jar jarfile.jar. Building excluding spark SQL is working fine

    – Binod Sarkar
    Nov 25 '18 at 4:12








  • 1





    Use spark-submit to run your jar. Follow the link: spark.apache.org/docs/latest/submitting-applications.html

    – anuj saxena
    Nov 25 '18 at 4:17











  • I have tried the spark-submit also but same error

    – Binod Sarkar
    Nov 25 '18 at 4:21














0












0








0








I am new to big data. I have developed an spark application using Scala to communicate with hive. It is working fine on Intellij idea. But when I am building an jar file with all dependency I am getting the error




Error: Could not find or load main class




After further debugging I have found that when I am including spark SQL or hive then I am getting the error. It is a maven project. Please have a look on my pom.xml



<modelVersion>4.0.0</modelVersion>

<groupId>spk</groupId>
<artifactId>spk</artifactId>
<version>1.0-SNAPSHOT</version>

<dependencies>
<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-core -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-sql -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!--<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive_2.11</artifactId>
<version>2.0.2</version>
</dependency>-->

<!--<dependency>
<groupId>org.apache.hive</groupId>
<artifactId>hive-exec</artifactId>
<version>1.2.2</version>
</dependency>-->

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka_2.11</artifactId>
<version>2.0.1</version>
</dependency>

<!-- https://mvnrepository.com/artifact/com.101tec/zkclient -->
<dependency>
<groupId>com.101tec</groupId>
<artifactId>zkclient</artifactId>
<version>0.11</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka-clients</artifactId>
<version>0.10.0.0</version>
</dependency>

</dependencies>

<build>
<sourceDirectory>src/main/scala</sourceDirectory>
<plugins>
<plugin>
<groupId>org.scala-tools</groupId>
<artifactId>maven-scala-plugin</artifactId>
<version>2.11</version>
<executions>
<execution>
<goals>
<goal>compile</goal>
<goal>testCompile</goal>
</goals>
<configuration>
<args>
<!--<arg>-make:transitive</arg>-->
<arg>-dependencyfile</arg>
<arg>${project.build.directory}/.scala_dependencies</arg>
</args>
</configuration>
</execution>
</executions>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-surefire-plugin</artifactId>
<version>2.11</version>
<configuration>
<useFile>false</useFile>
<disableXmlReport>true</disableXmlReport>
<!-- If you have classpath issue like NoDefClassError,... -->
<!-- useManifestOnlyJar>false</useManifestOnlyJar -->
<includes>
<include>**/*Test.*</include>
<include>**/*Suite.*</include>
</includes>
</configuration>
</plugin>
</plugins>
</build>


Thanks in advance.










share|improve this question














I am new to big data. I have developed an spark application using Scala to communicate with hive. It is working fine on Intellij idea. But when I am building an jar file with all dependency I am getting the error




Error: Could not find or load main class




After further debugging I have found that when I am including spark SQL or hive then I am getting the error. It is a maven project. Please have a look on my pom.xml



<modelVersion>4.0.0</modelVersion>

<groupId>spk</groupId>
<artifactId>spk</artifactId>
<version>1.0-SNAPSHOT</version>

<dependencies>
<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-core -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.spark/spark-sql -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_2.11</artifactId>
<version>2.0.2</version>
</dependency>

<!--<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive_2.11</artifactId>
<version>2.0.2</version>
</dependency>-->

<!--<dependency>
<groupId>org.apache.hive</groupId>
<artifactId>hive-exec</artifactId>
<version>1.2.2</version>
</dependency>-->

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka_2.11</artifactId>
<version>2.0.1</version>
</dependency>

<!-- https://mvnrepository.com/artifact/com.101tec/zkclient -->
<dependency>
<groupId>com.101tec</groupId>
<artifactId>zkclient</artifactId>
<version>0.11</version>
</dependency>

<!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka-clients</artifactId>
<version>0.10.0.0</version>
</dependency>

</dependencies>

<build>
<sourceDirectory>src/main/scala</sourceDirectory>
<plugins>
<plugin>
<groupId>org.scala-tools</groupId>
<artifactId>maven-scala-plugin</artifactId>
<version>2.11</version>
<executions>
<execution>
<goals>
<goal>compile</goal>
<goal>testCompile</goal>
</goals>
<configuration>
<args>
<!--<arg>-make:transitive</arg>-->
<arg>-dependencyfile</arg>
<arg>${project.build.directory}/.scala_dependencies</arg>
</args>
</configuration>
</execution>
</executions>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-surefire-plugin</artifactId>
<version>2.11</version>
<configuration>
<useFile>false</useFile>
<disableXmlReport>true</disableXmlReport>
<!-- If you have classpath issue like NoDefClassError,... -->
<!-- useManifestOnlyJar>false</useManifestOnlyJar -->
<includes>
<include>**/*Test.*</include>
<include>**/*Suite.*</include>
</includes>
</configuration>
</plugin>
</plugins>
</build>


Thanks in advance.







scala maven apache-spark hive






share|improve this question













share|improve this question











share|improve this question




share|improve this question










asked Nov 25 '18 at 3:48









Binod SarkarBinod Sarkar

14




14








  • 1





    What is the command you are using to run your jar?

    – anuj saxena
    Nov 25 '18 at 4:02











  • java -jar jarfile.jar. Building excluding spark SQL is working fine

    – Binod Sarkar
    Nov 25 '18 at 4:12








  • 1





    Use spark-submit to run your jar. Follow the link: spark.apache.org/docs/latest/submitting-applications.html

    – anuj saxena
    Nov 25 '18 at 4:17











  • I have tried the spark-submit also but same error

    – Binod Sarkar
    Nov 25 '18 at 4:21














  • 1





    What is the command you are using to run your jar?

    – anuj saxena
    Nov 25 '18 at 4:02











  • java -jar jarfile.jar. Building excluding spark SQL is working fine

    – Binod Sarkar
    Nov 25 '18 at 4:12








  • 1





    Use spark-submit to run your jar. Follow the link: spark.apache.org/docs/latest/submitting-applications.html

    – anuj saxena
    Nov 25 '18 at 4:17











  • I have tried the spark-submit also but same error

    – Binod Sarkar
    Nov 25 '18 at 4:21








1




1





What is the command you are using to run your jar?

– anuj saxena
Nov 25 '18 at 4:02





What is the command you are using to run your jar?

– anuj saxena
Nov 25 '18 at 4:02













java -jar jarfile.jar. Building excluding spark SQL is working fine

– Binod Sarkar
Nov 25 '18 at 4:12







java -jar jarfile.jar. Building excluding spark SQL is working fine

– Binod Sarkar
Nov 25 '18 at 4:12






1




1





Use spark-submit to run your jar. Follow the link: spark.apache.org/docs/latest/submitting-applications.html

– anuj saxena
Nov 25 '18 at 4:17





Use spark-submit to run your jar. Follow the link: spark.apache.org/docs/latest/submitting-applications.html

– anuj saxena
Nov 25 '18 at 4:17













I have tried the spark-submit also but same error

– Binod Sarkar
Nov 25 '18 at 4:21





I have tried the spark-submit also but same error

– Binod Sarkar
Nov 25 '18 at 4:21












1 Answer
1






active

oldest

votes


















0














Thanks for your help. I have got the answer. There are some problem with Intellij Idea build with all dependency.
I have done the following steps:




1) Added to the pom.xml




<plugin>
<artifactId>maven-assembly-plugin</artifactId>
<version>3.1.0</version>
<configuration>
<archive>
<manifest>
<mainClass>spk.kspark</mainClass>
</manifest>
</archive>
<descriptorRefs>
<descriptorRef>jar-with-dependencies</descriptorRef>
</descriptorRefs>
</configuration>
</plugin>



2) Go to the project directory and open Ubuntu terminal



3) mvn install



4) mvn clean compile assembly:single or mvn package assembly:single







share|improve this answer

























    Your Answer






    StackExchange.ifUsing("editor", function () {
    StackExchange.using("externalEditor", function () {
    StackExchange.using("snippets", function () {
    StackExchange.snippets.init();
    });
    });
    }, "code-snippets");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "1"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53464492%2fspark-sql-and-hive-causing-error-could-not-find-or-load-main-class%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    0














    Thanks for your help. I have got the answer. There are some problem with Intellij Idea build with all dependency.
    I have done the following steps:




    1) Added to the pom.xml




    <plugin>
    <artifactId>maven-assembly-plugin</artifactId>
    <version>3.1.0</version>
    <configuration>
    <archive>
    <manifest>
    <mainClass>spk.kspark</mainClass>
    </manifest>
    </archive>
    <descriptorRefs>
    <descriptorRef>jar-with-dependencies</descriptorRef>
    </descriptorRefs>
    </configuration>
    </plugin>



    2) Go to the project directory and open Ubuntu terminal



    3) mvn install



    4) mvn clean compile assembly:single or mvn package assembly:single







    share|improve this answer






























      0














      Thanks for your help. I have got the answer. There are some problem with Intellij Idea build with all dependency.
      I have done the following steps:




      1) Added to the pom.xml




      <plugin>
      <artifactId>maven-assembly-plugin</artifactId>
      <version>3.1.0</version>
      <configuration>
      <archive>
      <manifest>
      <mainClass>spk.kspark</mainClass>
      </manifest>
      </archive>
      <descriptorRefs>
      <descriptorRef>jar-with-dependencies</descriptorRef>
      </descriptorRefs>
      </configuration>
      </plugin>



      2) Go to the project directory and open Ubuntu terminal



      3) mvn install



      4) mvn clean compile assembly:single or mvn package assembly:single







      share|improve this answer




























        0












        0








        0







        Thanks for your help. I have got the answer. There are some problem with Intellij Idea build with all dependency.
        I have done the following steps:




        1) Added to the pom.xml




        <plugin>
        <artifactId>maven-assembly-plugin</artifactId>
        <version>3.1.0</version>
        <configuration>
        <archive>
        <manifest>
        <mainClass>spk.kspark</mainClass>
        </manifest>
        </archive>
        <descriptorRefs>
        <descriptorRef>jar-with-dependencies</descriptorRef>
        </descriptorRefs>
        </configuration>
        </plugin>



        2) Go to the project directory and open Ubuntu terminal



        3) mvn install



        4) mvn clean compile assembly:single or mvn package assembly:single







        share|improve this answer















        Thanks for your help. I have got the answer. There are some problem with Intellij Idea build with all dependency.
        I have done the following steps:




        1) Added to the pom.xml




        <plugin>
        <artifactId>maven-assembly-plugin</artifactId>
        <version>3.1.0</version>
        <configuration>
        <archive>
        <manifest>
        <mainClass>spk.kspark</mainClass>
        </manifest>
        </archive>
        <descriptorRefs>
        <descriptorRef>jar-with-dependencies</descriptorRef>
        </descriptorRefs>
        </configuration>
        </plugin>



        2) Go to the project directory and open Ubuntu terminal



        3) mvn install



        4) mvn clean compile assembly:single or mvn package assembly:single








        share|improve this answer














        share|improve this answer



        share|improve this answer








        edited Nov 26 '18 at 8:45

























        answered Nov 25 '18 at 7:12









        Binod SarkarBinod Sarkar

        14




        14






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Stack Overflow!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53464492%2fspark-sql-and-hive-causing-error-could-not-find-or-load-main-class%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Contact image not getting when fetch all contact list from iPhone by CNContact

            count number of partitions of a set with n elements into k subsets

            A CLEAN and SIMPLE way to add appendices to Table of Contents and bookmarks