Hot-keys on this page
r m x p toggle line displays
j k next/prev highlighted chunk
0 (zero) top of page
1 (one) first highlighted chunk
# # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
PySpark is the Python API for Spark.
Public classes:
- :class:`SparkContext`: Main entry point for Spark functionality. - :class:`RDD`: A Resilient Distributed Dataset (RDD), the basic abstraction in Spark. - :class:`Broadcast`: A broadcast variable that gets reused across tasks. - :class:`Accumulator`: An "add-only" shared variable that tasks can only add values to. - :class:`SparkConf`: For configuring Spark. - :class:`SparkFiles`: Access files shipped with jobs. - :class:`StorageLevel`: Finer-grained cache persistence levels. - :class:`TaskContext`: Information about the current running task, available on the workers and experimental. - :class:`RDDBarrier`: Wraps an RDD under a barrier stage for barrier execution. - :class:`BarrierTaskContext`: A :class:`TaskContext` that provides extra info and tooling for barrier execution. - :class:`BarrierTaskInfo`: Information about a barrier task. - :class:`InheritableThread`: A inheritable thread to use in Spark when the pinned thread mode is on. """
""" A decorator that annotates a function to append the version of Spark the function was added. """
""" Returns a function with same code, globals, defaults, closure, and name (or provide a new name). """ # See # http://stackoverflow.com/questions/6527633/how-can-i-make-a-deepcopy-of-a-function-in-python f.__closure__) # in case f was given attrs (note this dict is a shallow copy):
""" A decorator that forces keyword arguments in the wrapped method and saves actual input keyword arguments in `_input_kwargs`.
Notes ----- Should only be used to wrap a method where first arg is `self` """ def wrapper(self, *args, **kwargs):
# To avoid circular dependencies
# for back compatibility
"SparkConf", "SparkContext", "SparkFiles", "RDD", "StorageLevel", "Broadcast", "Accumulator", "AccumulatorParam", "MarshalSerializer", "PickleSerializer", "StatusTracker", "SparkJobInfo", "SparkStageInfo", "Profiler", "BasicProfiler", "TaskContext", "RDDBarrier", "BarrierTaskContext", "BarrierTaskInfo", "InheritableThread", "inheritable_thread_target", "__version__", ] |