Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions python/docs/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@ Contents:
pyspark.streaming
pyspark.ml
pyspark.mllib
pyspark.resource


Core classes:
Expand Down
11 changes: 11 additions & 0 deletions python/docs/pyspark.resource.rst
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
pyspark.resource module
=======================

Module Contents
---------------

.. automodule:: pyspark.resource
:members:
:undoc-members:
:inherited-members:

1 change: 1 addition & 0 deletions python/docs/pyspark.rst
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,7 @@ Subpackages
pyspark.streaming
pyspark.ml
pyspark.mllib
pyspark.resource

Contents
--------
Expand Down
4 changes: 2 additions & 2 deletions python/docs/pyspark.sql.rst
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
pyspark.sql module
==================

Module Context
--------------
Module Contents
---------------

.. automodule:: pyspark.sql
:members:
Expand Down
2 changes: 1 addition & 1 deletion python/pyspark/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,7 @@
from pyspark.storagelevel import StorageLevel
from pyspark.accumulators import Accumulator, AccumulatorParam
from pyspark.broadcast import Broadcast
from pyspark.resourceinformation import ResourceInformation
from pyspark.resource import ResourceInformation
from pyspark.serializers import MarshalSerializer, PickleSerializer
from pyspark.status import *
from pyspark.taskcontext import TaskContext, BarrierTaskContext, BarrierTaskInfo
Expand Down
2 changes: 1 addition & 1 deletion python/pyspark/context.py
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@
from pyspark.serializers import PickleSerializer, BatchedSerializer, UTF8Deserializer, \
PairDeserializer, AutoBatchedSerializer, NoOpSerializer, ChunkedStream
from pyspark.storagelevel import StorageLevel
from pyspark.resourceinformation import ResourceInformation
from pyspark.resource import ResourceInformation
from pyspark.rdd import RDD, _load_from_socket, ignore_unicode_prefix
from pyspark.traceback_utils import CallSite, first_spark_call
from pyspark.status import StatusTracker
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,8 +26,10 @@ class ResourceInformation(object):

One example is GPUs, where the addresses would be the indices of the GPUs

@param name the name of the resource
@param addresses an array of strings describing the addresses of the resource
:param name: the name of the resource
:param addresses: an array of strings describing the addresses of the resource

.. versionadded:: 3.0.0
"""

def __init__(self, name, addresses):
Expand Down
3 changes: 2 additions & 1 deletion python/pyspark/worker.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,7 @@
Worker that receives input from Piped RDD.
"""
from __future__ import print_function
from __future__ import absolute_import
import os
import sys
import time
Expand All @@ -35,7 +36,7 @@
from pyspark.java_gateway import local_connect_and_auth
from pyspark.taskcontext import BarrierTaskContext, TaskContext
from pyspark.files import SparkFiles
from pyspark.resourceinformation import ResourceInformation
from pyspark.resource import ResourceInformation
from pyspark.rdd import PythonEvalType
from pyspark.serializers import write_with_length, write_int, read_long, read_bool, \
write_long, read_int, SpecialLengths, UTF8Deserializer, PickleSerializer, \
Expand Down