Skip to content

Commit c318b24

Browse files
committed
Use CSS styles
1 parent e4c76b9 commit c318b24

35 files changed

+95
-95
lines changed

core/src/main/scala/org/apache/spark/Aggregator.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ package org.apache.spark
2020
import org.apache.spark.util.collection.{AppendOnlyMap, ExternalAppendOnlyMap}
2121

2222
/**
23-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
23+
* <span class="developer badge">Developer API</span>
2424
* A set of functions used to aggregate data.
2525
*
2626
* @param createCombiner function to create the initial value of the aggregation.

core/src/main/scala/org/apache/spark/Dependency.scala

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -21,14 +21,14 @@ import org.apache.spark.rdd.RDD
2121
import org.apache.spark.serializer.Serializer
2222

2323
/**
24-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
24+
* <span class="developer badge">Developer API</span>
2525
* Base class for dependencies.
2626
*/
2727
abstract class Dependency[T](val rdd: RDD[T]) extends Serializable
2828

2929

3030
/**
31-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
31+
* <span class="developer badge">Developer API</span>
3232
* Base class for dependencies where each partition of the parent RDD is used by at most one
3333
* partition of the child RDD. Narrow dependencies allow for pipelined execution.
3434
*/
@@ -43,7 +43,7 @@ abstract class NarrowDependency[T](rdd: RDD[T]) extends Dependency(rdd) {
4343

4444

4545
/**
46-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
46+
* <span class="developer badge">Developer API</span>
4747
* Represents a dependency on the output of a shuffle stage.
4848
* @param rdd the parent RDD
4949
* @param partitioner partitioner used to partition the shuffle output
@@ -62,7 +62,7 @@ class ShuffleDependency[K, V](
6262

6363

6464
/**
65-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
65+
* <span class="developer badge">Developer API</span>
6666
* Represents a one-to-one dependency between partitions of the parent and child RDDs.
6767
*/
6868
class OneToOneDependency[T](rdd: RDD[T]) extends NarrowDependency[T](rdd) {
@@ -71,7 +71,7 @@ class OneToOneDependency[T](rdd: RDD[T]) extends NarrowDependency[T](rdd) {
7171

7272

7373
/**
74-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
74+
* <span class="developer badge">Developer API</span>
7575
* Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs.
7676
* @param rdd the parent RDD
7777
* @param inStart the start of the range in the parent RDD

core/src/main/scala/org/apache/spark/FutureAction.scala

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ import org.apache.spark.rdd.RDD
2525
import org.apache.spark.scheduler.{JobFailed, JobSucceeded, JobWaiter}
2626

2727
/**
28-
* <span class="badge" style="float: right; background-color: #257080;">EXPERIMENTAL</span>
28+
* <span class="experimental badge">Experimental</span>
2929
* A future for the result of an action to support cancellation. This is an extension of the
3030
* Scala Future interface to support cancellation.
3131
*/
@@ -85,7 +85,7 @@ trait FutureAction[T] extends Future[T] {
8585

8686

8787
/**
88-
* <span class="badge" style="float: right; background-color: #257080;">EXPERIMENTAL</span>
88+
* <span class="experimental badge">Experimental</span>
8989
* A [[FutureAction]] holding the result of an action that triggers a single job. Examples include
9090
* count, collect, reduce.
9191
*/
@@ -150,7 +150,7 @@ class SimpleFutureAction[T] private[spark](jobWaiter: JobWaiter[_], resultFunc:
150150

151151

152152
/**
153-
* <span class="badge" style="float: right; background-color: #257080;">EXPERIMENTAL</span>
153+
* <span class="experimental badge">Experimental</span>
154154
* A [[FutureAction]] for actions that could trigger multiple Spark jobs. Examples include take,
155155
* takeSample. Cancellation works by setting the cancelled flag to true and interrupting the
156156
* action thread if it is being blocked by a job.

core/src/main/scala/org/apache/spark/Logging.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ import org.slf4j.{Logger, LoggerFactory}
2222
import org.slf4j.impl.StaticLoggerBinder
2323

2424
/**
25-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
25+
* <span class="developer badge">Developer API</span>
2626
* Utility trait for classes that want to log data. Creates a SLF4J logger for the class and allows
2727
* logging messages at different levels using methods that only evaluate parameters lazily if the
2828
* log level is enabled.

core/src/main/scala/org/apache/spark/SerializableWritable.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ import org.apache.hadoop.conf.Configuration
2323
import org.apache.hadoop.io.ObjectWritable
2424
import org.apache.hadoop.io.Writable
2525

26-
/** <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span> */
26+
/** <span class="developer badge">Developer API</span> */
2727
class SerializableWritable[T <: Writable](@transient var t: T) extends Serializable {
2828
def value = t
2929
override def toString = t.toString

core/src/main/scala/org/apache/spark/SparkContext.scala

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -63,13 +63,13 @@ class SparkContext(config: SparkConf)
6363
private[spark] var preferredNodeLocationData: Map[String, Set[SplitInfo]] = Map()
6464

6565
/**
66-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
66+
* <span class="developer badge">Developer API</span>
6767
* Alternative constructor for setting preferred locations where Spark will create executors.
6868
*
6969
* @param preferredNodeLocationData used in YARN mode to select nodes to launch containers on. Ca
7070
* be generated using [[org.apache.spark.scheduler.InputFormatInfo.computePreferredLocations]]
7171
* from a list of input files or InputFormats for the application.
72-
*/
72+
7373
def this(config: SparkConf, preferredNodeLocationData: Map[String, Set[SplitInfo]]) = {
7474
this(config)
7575
this.preferredNodeLocationData = preferredNodeLocationData
@@ -714,7 +714,7 @@ class SparkContext(config: SparkConf)
714714
}
715715
716716
/**
717-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
717+
* <span class="developer badge">Developer API</span>
718718
* Register a listener to receive up-calls from events that happen during execution.
719719
*/
720720
def addSparkListener(listener: SparkListener) {
@@ -1026,7 +1026,7 @@ class SparkContext(config: SparkConf)
10261026
}
10271027
10281028
/**
1029-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
1029+
* <span class="developer badge">Developer API</span>
10301030
* Run a job that can return approximate results.
10311031
*/
10321032
def runApproximateJob[T, U, R](
@@ -1044,7 +1044,7 @@ class SparkContext(config: SparkConf)
10441044
}
10451045
10461046
/**
1047-
* <span class="badge" style="float: right; background-color: #257080;">EXPERIMENTAL</span>
1047+
* <span class="experimental badge">Experimental</span>
10481048
* Submit a job for execution and return a FutureJob holding the result.
10491049
*/
10501050
def submitJob[T, U, R](

core/src/main/scala/org/apache/spark/SparkEnv.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ import org.apache.spark.storage._
3535
import org.apache.spark.util.{AkkaUtils, Utils}
3636

3737
/**
38-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
38+
* <span class="developer badge">Developer API</span>
3939
* Holds all the runtime environment objects for a running Spark instance (either master or worker),
4040
* including the serializer, Akka actor system, block manager, map output tracker, etc. Currently
4141
* Spark code finds the SparkEnv through a thread-local variable, so each thread that accesses these

core/src/main/scala/org/apache/spark/TaskContext.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ import scala.collection.mutable.ArrayBuffer
2222
import org.apache.spark.executor.TaskMetrics
2323

2424
/**
25-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
25+
* <span class="developer badge">Developer API</span>
2626
* Contextual information about a task which can be read or mutated during execution.
2727
*/
2828
class TaskContext(

core/src/main/scala/org/apache/spark/broadcast/BroadcastFactory.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ import org.apache.spark.SecurityManager
2121
import org.apache.spark.SparkConf
2222

2323
/**
24-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
24+
* <span class="developer badge">Developer API</span>
2525
* An interface for all the broadcast implementations in Spark (to allow
2626
* multiple broadcast implementations). SparkContext uses a user-specified
2727
* BroadcastFactory implementation to instantiate a particular broadcast for the

core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ package org.apache.spark.executor
2020
import org.apache.spark.storage.{BlockId, BlockStatus}
2121

2222
/**
23-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
23+
* <span class="developer badge">Developer API</span>
2424
* Metrics tracked during the execution of a task.
2525
*/
2626
class TaskMetrics extends Serializable {
@@ -87,7 +87,7 @@ object TaskMetrics {
8787

8888

8989
/**
90-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
90+
* <span class="developer badge">Developer API</span>
9191
* Metrics pertaining to shuffle data read in a given task.
9292
*/
9393
class ShuffleReadMetrics extends Serializable {
@@ -125,7 +125,7 @@ class ShuffleReadMetrics extends Serializable {
125125
}
126126

127127
/**
128-
* <span class="badge" style="float: right; background-color: #44751E;">DEVELOPER API</span>
128+
* <span class="developer badge">Developer API</span>
129129
* Metrics pertaining to shuffle data written in a given task.
130130
*/
131131
class ShuffleWriteMetrics extends Serializable {

0 commit comments

Comments
 (0)