[Spark-Study] Day-4 스파크 로컬 디버깅

2021. 8. 5. 09:56BigDATA/spark

반응형

2021.06.14 - [Study/Study group] - [Spark-Study] Day-1

2021.06.24 - [Study/Study group] - [Spark-Study] Day-2

2021.07.01 - [BigDATA/spark] - [Spark-Study] Day-3

 

spark local debugging 해보기

코드상으로는 아래의 내용이 추가 된다.

  val conf = new SparkConf().setAppName("MnMCount")
    conf.setIfMissing("spark.master", "local[*]")

    val spark = SparkSession
      .builder
      .config(conf)
      .getOrCreate()

//    val spark = SparkSession
//      .builder
//      .appName("MnMCount")
//      .getOrCreate()

실행의 경우 Debugging용 Application을 하나 추가한다.

그리고나서 벌레모양 아이콘을 꾹~눌러준다

 

아래처럼 브레이크 포인트가 잡히는걸 볼수가 있다.

 

참고 사이트 : https://supergloo.com/spark-scala/debug-scala-spark-intellij/

  1. Local Spark Debugging
  2. Remote Spark Debugging

챕터 3 스터디(읽고 토론)

 

spark basic Data types

cd /usr/local/Cellar/apache-spark/3.1.2
 terrypark  /usr/local/Cellar/apache-spark/3.1.2
 ll
total 192
-rw-r--r--   1 terrypark  admin   1.3K  7 29 13:45 INSTALL_RECEIPT.json
-rw-r--r--   1 terrypark  admin    23K  5 24 13:45 LICENSE
-rw-r--r--   1 terrypark  admin    56K  5 24 13:45 NOTICE
-rw-r--r--   1 terrypark  admin   4.4K  5 24 13:45 README.md
drwxr-xr-x  13 terrypark  admin   416B  7 29 13:45 bin
drwxr-xr-x  14 terrypark  admin   448B  5 24 13:45 libexec
 terrypark  /usr/local/Cellar/apache-spark/3.1.2
 ./bin/spark-shell
21/08/05 10:38:03 WARN Utils: Your hostname, acetui-MacBookPro.local resolves to a loopback address: 127.0.0.1; using ip instead (on interface en0)
21/08/05 10:38:03 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform (file:/usr/local/Cellar/apache-spark/3.1.2/libexec/jars/spark-unsafe_2.12-3.1.2.jar) to constructor java.nio.DirectByteBuffer(long,int)
WARNING: Please consider reporting this to the maintainers of org.apache.spark.unsafe.Platform
WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
WARNING: All illegal access operations will be denied in a future release
21/08/05 10:38:03 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
21/08/05 10:38:08 WARN Utils: Service 'SparkUI' could not bind on port 4040. Attempting port 4041.
Spark context Web UI available at http://ip:4041
Spark context available as 'sc' (master = local[*], app id = local-1628127488293).
Spark session available as 'spark'.
Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 3.1.2
      /_/

Using Scala version 2.12.10 (OpenJDK 64-Bit Server VM, Java 11.0.10)
Type in expressions to have them evaluated.
Type :help for more information.

scala> import org.apache.spark.sql.types._
import org.apache.spark.sql.types._

scala> val nameTypes = StringType
nameTypes: org.apache.spark.sql.types.StringType.type = StringType

scala> val firstName = nameTypes
firstName: org.apache.spark.sql.types.StringType.type = StringType

scala> val lastName = nameTypes
lastName: org.apache.spark.sql.types.StringType.type = StringType

아래 테이블 참고! 데이터 타입~

위처럼 너무 복잡스러워서 Schema가 나온다.
스키마를 정의 하는데는 2가지 방법이 있다.

관련 github : https://github.com/databricks/LearningSparkV2

next study : 챕터 3 코드 작성 후 실행

 

반응형
LIST