|  | 
| 17 | 17 | 
 | 
| 18 | 18 | package org.apache.spark | 
| 19 | 19 | 
 | 
|  | 20 | +import scala.language.implicitConversions | 
|  | 21 | + | 
| 20 | 22 | import java.io._ | 
| 21 | 23 | import java.net.URI | 
| 22 |  | -import java.util.UUID.randomUUID | 
| 23 | 24 | import java.util.concurrent.atomic.AtomicInteger | 
| 24 | 25 | import java.util.{Properties, UUID} | 
| 25 |  | - | 
|  | 26 | +import java.util.UUID.randomUUID | 
|  | 27 | +import scala.collection.{Map, Set} | 
|  | 28 | +import scala.collection.JavaConversions._ | 
|  | 29 | +import scala.collection.generic.Growable | 
|  | 30 | +import scala.collection.mutable.HashMap | 
|  | 31 | +import scala.reflect.{ClassTag, classTag} | 
| 26 | 32 | import org.apache.hadoop.conf.Configuration | 
| 27 | 33 | import org.apache.hadoop.fs.Path | 
| 28 | 34 | import org.apache.hadoop.io.{ArrayWritable, BooleanWritable, BytesWritable, DoubleWritable, FloatWritable, IntWritable, LongWritable, NullWritable, Text, Writable} | 
| 29 | 35 | import org.apache.hadoop.mapred.{FileInputFormat, InputFormat, JobConf, SequenceFileInputFormat, TextInputFormat} | 
| 30 |  | -import org.apache.hadoop.mapreduce.lib.input.{FileInputFormat => NewFileInputFormat} | 
| 31 | 36 | import org.apache.hadoop.mapreduce.{InputFormat => NewInputFormat, Job => NewHadoopJob} | 
|  | 37 | +import org.apache.hadoop.mapreduce.lib.input.{FileInputFormat => NewFileInputFormat} | 
| 32 | 38 | import org.apache.mesos.MesosNativeLibrary | 
|  | 39 | + | 
| 33 | 40 | import org.apache.spark.annotation.{DeveloperApi, Experimental} | 
| 34 | 41 | import org.apache.spark.broadcast.Broadcast | 
| 35 | 42 | import org.apache.spark.deploy.{LocalSparkCluster, SparkHadoopUtil} | 
| 36 |  | -import org.apache.spark.input.{ByteInputFormat, FixedLengthBinaryInputFormat, StreamInputFormat, WholeTextFileInputFormat} | 
|  | 43 | +import org.apache.spark.input.{StreamInputFormat, StreamFileInputFormat, WholeTextFileInputFormat, ByteInputFormat, FixedLengthBinaryInputFormat} | 
| 37 | 44 | import org.apache.spark.partial.{ApproximateEvaluator, PartialResult} | 
| 38 | 45 | import org.apache.spark.rdd._ | 
| 39 | 46 | import org.apache.spark.scheduler._ | 
|  | 47 | +import org.apache.spark.scheduler.cluster.{CoarseGrainedSchedulerBackend, SparkDeploySchedulerBackend, SimrSchedulerBackend} | 
| 40 | 48 | import org.apache.spark.scheduler.cluster.mesos.{CoarseMesosSchedulerBackend, MesosSchedulerBackend} | 
| 41 |  | -import org.apache.spark.scheduler.cluster.{CoarseGrainedSchedulerBackend, SimrSchedulerBackend, SparkDeploySchedulerBackend} | 
| 42 | 49 | import org.apache.spark.scheduler.local.LocalBackend | 
| 43 | 50 | import org.apache.spark.storage.{BlockManagerSource, RDDInfo, StorageStatus, StorageUtils} | 
| 44 | 51 | import org.apache.spark.ui.SparkUI | 
| 45 | 52 | import org.apache.spark.util.{CallSite, ClosureCleaner, MetadataCleaner, MetadataCleanerType, TimeStampedWeakValueHashMap, Utils} | 
| 46 | 53 | 
 | 
| 47 |  | -import scala.collection.JavaConversions._ | 
| 48 |  | -import scala.collection.generic.Growable | 
| 49 |  | -import scala.collection.mutable.HashMap | 
| 50 |  | -import scala.collection.{Map, Set} | 
| 51 |  | -import scala.language.implicitConversions | 
| 52 |  | -import scala.reflect.{ClassTag, classTag} | 
| 53 |  | - | 
| 54 | 54 | /** | 
| 55 | 55 |  * Main entry point for Spark functionality. A SparkContext represents the connection to a Spark | 
| 56 | 56 |  * cluster, and can be used to create RDDs, accumulators and broadcast variables on that cluster. | 
|  | 
0 commit comments