package rdd.sparksql;

import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.SparkSession;

//对数据进行封装，实际上就是在rdd的基础上对rdd在封装一次，变成dataset，一行数据就是按照row对象进行存储
public class SparkSQL103_Dataset {
    public static void main(String[] args) {
        final SparkSession sparkSession = SparkSession.builder()
                .master("local[*]")
                .appName("SparkSQL")
                .getOrCreate();

        final Dataset<Row> ds = sparkSession.read().json("data/user.json");
    }
}
