赞
踩
打开IDEA,选择File->Setting,搜索Scala安装
进入IDEA-Scala插件,下载符合自己IDEA版本的插件,找到下图位置安装即可
新建Project
选择新建Maven项目
之后修改项目名称->Next->Finish
在Settings中将编译环境设置为java 8,然后apply->ok
pom.xml文件内容示例
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd"> <modelVersion>4.0.0</modelVersion> <groupId>org.example</groupId> <artifactId>test</artifactId> <version>1.0-SNAPSHOT</version> <inceptionYear>2008</inceptionYear> <properties> <scala.version>2.11.8</scala.version> <spark.version>2.4.6</spark.version> <hadoop.version>2.7.3</hadoop.version> </properties> <repositories> <repository> <id>scala-tools.org</id> <name>Scala-Tools Maven2 Repository</name> <url>http://scala-tools.org/repo-releases</url> </repository> </repositories> <pluginRepositories> <pluginRepository> <id>scala-tools.org</id> <name>Scala-Tools Maven2 Repository</name> <url>http://scala-tools.org/repo-releases</url> </pluginRepository> </pluginRepositories> <dependencies> <dependency> <groupId>org.scala-lang</groupId> <artifactId>scala-library</artifactId> <version>${scala.version}</version> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-core_2.11</artifactId> <version>${spark.version}</version> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-sql_2.11</artifactId> <version>${spark.version}</version> </dependency> <dependency> <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-hdfs</artifactId> <version>${hadoop.version}</version> </dependency> <dependency> <groupId>org.apache.hadoop</groupId> <artifactId>hadoop-common</artifactId> <version>${hadoop.version}</version> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-mllib_2.11</artifactId> <version>${spark.version}</version> </dependency> </dependencies> </project>
此时需要根据本地环境修改配置文件中的Spark,Scala,Hadoop版本
修改pom.xml后右键找到Maven->Reimport,等待下方进度条完成导入
注意:导入过程中若是出现了问题,可以到mvnrepository寻找匹配的版本
新建一个Scala Class
选择创建Object
测试代码如下:
package org.example import org.apache.spark.sql.SparkSession object hello { def main(args: Array[String]) { //构建Spark对象 val ss = SparkSession .builder() .master("local[*]") .appName("movie") .getOrCreate() val sc = ss.sparkContext val data_path = "file:/C:/Users/yeyu/Desktop/data/word.txt" val data = sc.textFile(data_path) data.map(_.split(",")).map(f => (f(0), f(1), f(2))).foreach(println(_)) } }
我在本地桌面创建了如下文本文件
右键运行可以看到输出结果
至此已经完成了Scala-Spark程序的本地编写和运行
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。