IDEA創建Spark開發環境


  1. 下面安裝的六步和通過Maven安裝Java項目的方式相同
    1. 如果剛安裝完IDEA,直接點擊[Create New Project]
    2. 左側選擇Maven,然后先勾選上部的[Create from archetype],然后右側選擇[maven-archetype-quickstart]
    3. 填寫GroupID、ArtifactId,點擊Next
    4. 配置一下maven的位置以及配置文件和本地倉庫位置
    5. 配置項目存儲的路徑,接着點擊Next即可
    6. 進入項目后在右下角選擇[Enable Auto-Import],這樣在修改配置文件時自動導入了 
  2. 導入 scala SDK
  3. 選擇導入的scala SDK的地址
  4. 構建自己需要的項目結構(便於管理)
  5. 將對應的目錄結構改成對應的源碼包
  6. 導入需要的Spark開發的依賴
    <repositories>
        <repository>
          <id>central</id>
          <name>aliyun maven</name>
          <url>http://maven.aliyun.com/nexus/content/groups/public/</url>
          <layout>default</layout>
          <!-- 是否開啟發布版構件下載 -->
          <releases>
            <enabled>true</enabled>
          </releases>
          <!-- 是否開啟快照版構件下載 -->
          <snapshots>
            <enabled>false</enabled>
          </snapshots>
        </repository>
    
        <repository>
          <id>cloudera</id>
          <url>https://repository.cloudera.com/artifactory/cloudera-repos/</url>
        </repository>
        <repository>
          <id>jboss</id>
          <url>http://repository.jboss.com/nexus/content/groups/public</url>
        </repository>
    </repositories>
    
    
    <properties>
            <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
            <maven.compiler.source>1.7</maven.compiler.source>
            <maven.compiler.target>1.7</maven.compiler.target>
            <hadoop.version>2.6.0-cdh5.7.6</hadoop.version>
            <spark.version>2.2.0</spark.version>
            <mysq.version>5.1.27</mysq.version>
            <hbase.version>1.2.0-cdh5.7.6</hbase.version>
            <uasparser.version>0.6.1</uasparser.version>
    </properties>
    
    
    <dependencies>
            <dependency>
                <groupId>org.apache.spark</groupId>
                <artifactId>spark-core_2.11</artifactId>
                <version>${spark.version}</version>
            </dependency>
            <dependency>
                <groupId>org.apache.spark</groupId>
                <artifactId>spark-sql_2.11</artifactId>
                <version>${spark.version}</version>
            </dependency>
            <dependency>
                <groupId>org.apache.spark</groupId>
                <artifactId>spark-streaming_2.11</artifactId>
                <version>${spark.version}</version>
            </dependency>
            <dependency>
                <groupId>org.apache.spark</groupId>
                <artifactId>spark-streaming-kafka-0-8_2.11</artifactId>
                <version>2.2.0</version>
            </dependency>
            <dependency>
                <groupId>org.apache.spark</groupId>
                <artifactId>spark-hive_2.11</artifactId>
                <version>${spark.version}</version>
            </dependency>
    
            <dependency>
                <groupId>org.apache.hadoop</groupId>
                <artifactId>hadoop-client</artifactId>
                <version>${hadoop.version}</version>
            </dependency>
    
            <dependency>
                <groupId>mysql</groupId>
                <artifactId>mysql-connector-java</artifactId>
                <version>${mysq.version}</version>
            </dependency>
    
            <dependency>
                <groupId>org.apache.hbase</groupId>
                <artifactId>hbase-server</artifactId>
                <version>${hbase.version}</version>
            </dependency>
            <dependency>
                <groupId>org.apache.hbase</groupId>
                <artifactId>hbase-hadoop2-compat</artifactId>
                <version>1.2.0-cdh5.7.6</version>
            </dependency>
            <dependency>
                <groupId>org.apache.hbase</groupId>
                <artifactId>hbase-client</artifactId>
                <version>1.2.0-cdh5.7.6</version>
            </dependency>
    
            <dependency>
                <groupId>cz.mallat.uasparser</groupId>
                <artifactId>uasparser</artifactId>
                <version>${uasparser.version}</version>
            </dependency>
    
            <!-- Spark MLlib依賴包-->
            <dependency>
                <groupId>org.apache.spark</groupId>
                <artifactId>spark-mllib_2.11</artifactId>
                <version>2.2.0</version>
            </dependency>
            <dependency>
                <groupId>org.scalanlp</groupId>
                <artifactId>breeze_2.11</artifactId>
                <version>0.13.1</version>
            </dependency>
    
    <!--        <dependency>
                <groupId>com.github.fommil.netlib</groupId>
                <artifactId>all</artifactId>
                <version>1.1.2</version>
            </dependency>-->
            <dependency>
                <groupId>org.jblas</groupId>
                <artifactId>jblas</artifactId>
                <version>1.2.3</version>
            </dependency>
    
            <dependency>
                <groupId>org.mongodb.spark</groupId>
                <artifactId>mongo-spark-connector_2.11</artifactId>
                <version>2.3.1</version>
            </dependency>
    
            <dependency>
                <groupId>redis.clients</groupId>
                <artifactId>jedis</artifactId>
                <version>2.8.0</version>
            </dependency>
    
            <dependency>
                <groupId>org.apache.kafka</groupId>
                <artifactId>kafka-clients</artifactId>
                <version>0.8.2.1</version>
            </dependency>
    
            <dependency>
                <groupId>com.alibaba</groupId>
                <artifactId>fastjson</artifactId>
                <version>1.2.31</version>
            </dependency>
    </dependencies>


免責聲明!

本站轉載的文章為個人學習借鑒使用,本站對版權不負任何法律責任。如果侵犯了您的隱私權益,請聯系本站郵箱yoyou2525@163.com刪除。



 
粵ICP備18138465號   © 2018-2025 CODEPRJ.COM