package com.yaxin

import scala.math.random
//import org.apache.spark._

object SparkPi {
//  def main(args: Array[String]) {
//    //val conf = new SparkConf().setAppName("Spark Pi").setMaster("spark://hadoop:7070").setJars(List("out\\artifacts\\sparkTest_jar\\sparkTest.jar"))
//    val spark = new SparkContext("spark://hadoop:7070", "Spark Pi", "F:\\soft\\spark\\spark-1.1.0-bin-hadoop2.4", List("out\\artifacts\\sparkTest_jar\\sparkTest.jar"))
//    //val spark = new SparkContext(conf)
//    val slices = if (args.length > 0) args(0).toInt else 2
//    val n = 100000 * slices
//    val count = spark.parallelize(1 to n, slices).map { i =>
//      val x = random * 2 - 1
//      val y = random * 2 - 1
//      if (x*x + y*y < 1) 1 else 0
//    }.reduce(_ + _)
//    println("Pi is roughly " + 4.0 * count / n)
//    spark.stop()
//  }
}