Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
540 commits
Select commit Hold shift + click to select a range
3a4609e
HOTFIX: Rolling back incorrect version change
pwendell Nov 28, 2014
854fade
[SPARK-4597] Use proper exception and reset variable in Utils.createT…
viirya Nov 29, 2014
e07dbd8
[DOCS][BUILD] Add instruction to use change-version-to-2.11.sh in 'Bu…
ueshin Nov 30, 2014
d324728
SPARK-2143 [WEB UI] Add Spark version to UI footer
srowen Nov 30, 2014
c899f03
[SPARK-4656][Doc] Typo in Programming Guide markdown
Lewuathe Dec 1, 2014
0f4dad4
[DOC] Fixes formatting typo in SQL programming guide
liancheng Dec 1, 2014
9b8a769
SPARK-2192 [BUILD] Examples Data Not in Binary Distribution
srowen Dec 1, 2014
67a2c13
[SPARK-4661][Core] Minor code and docs cleanup
zsxwing Dec 1, 2014
35bc338
Documentation: add description for repartitionAndSortWithinPartitions
msiddalingaiah Dec 1, 2014
9c9b4bd
[SPARK-4258][SQL][DOC] Documents spark.sql.parquet.filterPushdown
liancheng Dec 1, 2014
e0a6d36
[SQL] add @group tab in limit() and count()
Dec 1, 2014
f2bb90a
[SPARK-4358][SQL] Let BigDecimal do checking type compatibility
viirya Dec 1, 2014
5006aab
[SPARK-4650][SQL] Supporting multi column support in countDistinct fu…
ravipesala Dec 1, 2014
b39cfee
[SPARK-4658][SQL] Code documentation issue in DDL of datasource API
ravipesala Dec 1, 2014
31cf51b
[SQL] Minor fix for doc and comment
scwf Dec 1, 2014
e66f816
[SQL][DOC] Date type in SQL programming guide
adrian-wang Dec 1, 2014
445fc95
[SPARK-4529] [SQL] support view with column alias
adrian-wang Dec 2, 2014
3783e15
[SPARK-4611][MLlib] Implement the efficient vector norm
Dec 2, 2014
b97c27f
[SPARK-4686] Link to allowed master URLs is broken
kayousterhout Dec 2, 2014
1850d90
[SPARK-4536][SQL] Add sqrt and abs to Spark SQL DSL
sarutak Dec 2, 2014
aa3d369
[SPARK-4663][sql]add finally to avoid resource leak
baishuo Dec 2, 2014
06129cd
[SPARK-4676][SQL] JavaSchemaRDD.schema may throw NullType MatchError …
YanTangZhai Dec 2, 2014
97dc238
[SPARK-4593][SQL] Return null when denominator is 0
adrian-wang Dec 2, 2014
adc5d6f
[SPARK-4670] [SQL] wrong symbol for bitwise not
adrian-wang Dec 2, 2014
658fe8f
[SPARK-4695][SQL] Get result using executeCollect
scwf Dec 2, 2014
5e026a3
[Release] Translate unknown author names automatically
Dec 3, 2014
f1859fc
[SPARK-4672][GraphX]Perform checkpoint() on PartitionsRDD to shorten …
Dec 3, 2014
528cce8
[SPARK-4672][GraphX]Non-transient PartitionsRDDs will lead to StackOv…
Dec 3, 2014
667f7ff
[SPARK-4672][Core]Checkpoint() should clear f to shorten the serializ…
Dec 3, 2014
fb14bfd
[SPARK-4710] [mllib] Eliminate MLlib compilation warnings
jkbradley Dec 3, 2014
8ff7a28
[SPARK-4708][MLLib] Make k-mean runs two/three times faster with dens…
Dec 3, 2014
b63e941
[SPARK-4717][MLlib] Optimize BLAS library to avoid de-reference multi…
Dec 3, 2014
163fd78
SPARK-2624 add datanucleus jars to the container in yarn-cluster
Dec 3, 2014
614e686
[SPARK-4701] Typo in sbt/sbt
tsudukim Dec 3, 2014
1ee65b4
[SPARK-4715][Core] Make sure tryToAcquire won't return a negative value
zsxwing Dec 3, 2014
4a71e08
[SPARK-4642] Add description about spark.yarn.queue to running-on-YAR…
tsudukim Dec 3, 2014
38cb2c3
[HOT FIX] [YARN] Check whether `/lib` exists before listing its files
Dec 3, 2014
4793197
[SPARK-4552][SQL] Avoid exception when reading empty parquet data thr…
marmbrus Dec 3, 2014
6b6b779
[SPARK-4498][core] Don't transition ExecutorInfo to RUNNING until Dri…
markhamstra Dec 3, 2014
fe28ee2
[SPARK-4085] Propagate FetchFailedException when Spark fails to read …
rxin Dec 4, 2014
4259ca8
[SPARK-4711] [mllib] [docs] Programming guide advice on choosing opti…
jkbradley Dec 4, 2014
9880bb4
[SPARK-4580] [SPARK-4610] [mllib] [docs] Documentation for tree ensem…
jkbradley Dec 4, 2014
f9e1f89
[Release] Correctly translate contributors name in release notes
Dec 4, 2014
2605acb
[SPARK-4685] Include all spark.ml and spark.mllib packages in JavaDoc…
Lewuathe Dec 4, 2014
dec838b
[SQL] Minor: Avoid calling Seq#size in a loop
aarondav Dec 4, 2014
bf720ef
[docs] Fix outdated comment in tuning guide
jkbradley Dec 4, 2014
266a814
[SPARK-4575] [mllib] [docs] spark.ml pipelines doc + bug fixes
jkbradley Dec 4, 2014
34fdca0
[FIX][DOC] Fix broken links in ml-guide.md
mengxr Dec 4, 2014
2fbe488
[SPARK-4683][SQL] Add a beeline.cmd to run on Windows
liancheng Dec 4, 2014
2c6e287
Revert "HOTFIX: Rolling back incorrect version change"
pwendell Dec 4, 2014
701019b
Revert "Preparing development version 1.2.1-SNAPSHOT"
pwendell Dec 4, 2014
078894c
Revert "Preparing Spark release v1.2.0-rc1"
pwendell Dec 4, 2014
d9aee07
[SPARK-4253] Ignore spark.driver.host in yarn-cluster and standalone-…
WangTaoTheTonic Dec 4, 2014
ead01b6
[HOTFIX] Fixing two issues with the release script.
pwendell Dec 4, 2014
2b72c56
Preparing Spark release v1.2.0-rc2
pwendell Dec 4, 2014
bc05df8
Preparing development version 1.2.1-SNAPSHOT
pwendell Dec 4, 2014
a00d0aa
[SPARK-4745] Fix get_existing_cluster() function with multiple securi…
alexdebrie Dec 4, 2014
0d159de
[SPARK-4459] Change groupBy type parameter from K to U
Dec 4, 2014
f5c5647
[SPARK-4652][DOCS] Add docs about spark-git-repo option
Lewuathe Dec 4, 2014
b905e11
[SPARK-4421] Wrong link in spark-standalone.html
tsudukim Dec 5, 2014
63b1bc1
Fix typo in Spark SQL docs.
andyk Dec 5, 2014
6c43631
[SPARK-4464] Description about configuration options need to be modif…
tsudukim Dec 5, 2014
325babe
Revert "[HOT FIX] [YARN] Check whether `/lib` exists before listing i…
Dec 5, 2014
a8d8077
Revert "SPARK-2624 add datanucleus jars to the container in yarn-clus…
Dec 5, 2014
d12ea49
[SPARK-4753][SQL] Use catalyst for partition pruning in newParquet.
marmbrus Dec 5, 2014
e8d8077
[SPARK-4761][SQL] Enables Kryo by default in Spark SQL Thrift server
liancheng Dec 5, 2014
11446a6
Streaming doc : do you mean inadvertently?
CrazyJvm Dec 5, 2014
27d9f13
[SPARK-3623][GraphX] GraphX should support the checkpoint operation
witgo Dec 6, 2014
a4ae7c8
[SPARK-4646] Replace Scala.util.Sorting.quickSort with Sorter(TimSort…
maropu Dec 8, 2014
6b9e8b0
[SPARK-4620] Add unpersist in Graph and GraphImpl
maropu Dec 8, 2014
9ed5641
[SPARK-4774] [SQL] Makes HiveFromSpark more portable
Dec 8, 2014
f416032
SPARK-4770. [DOC] [YARN] spark.scheduler.minRegisteredResourcesRatio …
sryza Dec 9, 2014
31a6d4f
[SPARK-4769] [SQL] CTAS does not work when reading from temporary tables
chenghao-intel Dec 9, 2014
e686742
[SPARK-4785][SQL] Initilize Hive UDFs on the driver and serialize the…
chenghao-intel Dec 9, 2014
5a3a3cc
[SPARK-4765] Make GC time always shown in UI.
kayousterhout Dec 9, 2014
51da2c5
SPARK-4567. Make SparkJobInfo and SparkStageInfo serializable
sryza Dec 10, 2014
b0d64e5
SPARK-4805 [CORE] BlockTransferMessage.toByteArray() trips assertion
srowen Dec 10, 2014
441ec34
[SPARK-4740] Create multiple concurrent connections between two peer …
rxin Dec 10, 2014
5e5d8f4
Config updates for the new shuffle transport.
rxin Dec 10, 2014
ff6f59b
[Minor] Use <sup> tag for help icon in web UI page header
JoshRosen Dec 10, 2014
a4d4a97
Revert "Preparing development version 1.2.1-SNAPSHOT"
pwendell Dec 10, 2014
e4f20bd
Revert "Preparing Spark release v1.2.0-rc2"
pwendell Dec 10, 2014
a428c44
Preparing Spark release v1.2.0-rc2
pwendell Dec 10, 2014
d70c729
Preparing development version 1.2.1-SNAPSHOT
pwendell Dec 10, 2014
1da1937
[SPARK-4771][Docs] Document standalone cluster supervise mode
Dec 10, 2014
1eb3ec5
SPARK-3526 Add section about data locality to the tuning guide
ash211 Dec 10, 2014
c3b0713
[SPARK-4806] Streaming doc update for 1.2
tdas Dec 11, 2014
c82e99d
[SPARK-4825] [SQL] CTAS fails to resolve when created using saveAsTable
chenghao-intel Dec 12, 2014
6eec4bc
fixed spelling errors in documentation
peterklipfel Dec 14, 2014
2ec78a1
HOTFIX: Disabling failing block manager test
pwendell Dec 15, 2014
c5a9ae6
[SPARK-4826] Fix generation of temp file names in WAL tests
JoshRosen Dec 15, 2014
ec19175
[SPARK-4668] Fix some documentation typos.
ryan-williams Dec 15, 2014
f1f27ec
[Minor][Core] fix comments in MapOutputTracker
scwf Dec 16, 2014
6bd8a96
SPARK-4814 [CORE] Enable assertions in SBT, Maven tests / AssertionEr…
srowen Dec 16, 2014
4f9916f
[DOCS][SQL] Add a Note on jsonFile having separate JSON objects per line
petervandenabeele Dec 16, 2014
1b6fc23
[SPARK-4847][SQL]Fix "extraStrategies cannot take effect in SQLContex…
jerryshao Dec 16, 2014
0fb0047
[Release] Major improvements to generate contributors script
Dec 17, 2014
8a69ed3
[Release] Cache known author translations locally
Dec 17, 2014
beb75ac
[Release] Update contributors list format and sort it
Dec 17, 2014
b5919d1
[HOTFIX] Fix RAT exclusion for known_translations file
JoshRosen Dec 17, 2014
2f00a29
[SPARK-4595][Core] Fix MetricsServlet not work issue
jerryshao Dec 17, 2014
e1d839e
[SPARK-4764] Ensure that files are fetched atomically
Dec 8, 2014
7ecf30e
[SPARK-4750] Dynamic allocation - synchronize kills
Dec 9, 2014
26dfac6
SPARK-3926 [CORE] Reopened: result of JavaRDD collectAsMap() is not s…
srowen Dec 9, 2014
51081e4
[SPARK-4691][shuffle] Restructure a few lines in shuffle code
Dec 9, 2014
0ebbccb
[SPARK-4714] BlockManager.dropFromMemory() should check whether block…
suyanNone Dec 9, 2014
e635168
[SPARK-4772] Clear local copies of accumulators as soon as we're done…
Dec 10, 2014
76c88c6
SPARK-785 [CORE] ClosureCleaner not invoked on most PairRDDFunctions
srowen Dec 16, 2014
0429ec3
[SPARK-4841] fix zip with textFile()
Dec 16, 2014
f305e7d
[SPARK-4821] [mllib] [python] [docs] Fix for pyspark.mllib.rand doc
jkbradley Dec 17, 2014
19efa5b
Add mesos specific configurations into doc
tnachen Dec 18, 2014
ef5c236
HOTFIX: Changing doc color
pwendell Dec 18, 2014
e7f9dd5
[SPARK-4880] remove spark.locality.wait in Analytics
Earne Dec 18, 2014
61c9b89
[SPARK-4884]: Improve Partition docs
msiddalingaiah Dec 19, 2014
075b399
[SPARK-4837] NettyBlockTransferService should use spark.blockManager.…
aarondav Dec 19, 2014
ca37639
[SPARK-4754] Refactor SparkContext into ExecutorAllocationClient
Dec 19, 2014
fd7bb9d
SPARK-3428. TaskMetrics for running tasks is missing GC time metrics
sryza Dec 19, 2014
6aa88cc
[SPARK-4889] update history server example cmds
ryan-williams Dec 19, 2014
f930fe8
[SPARK-4896] don’t redundantly overwrite executor JAR deps
ryan-williams Dec 19, 2014
4da1039
change signature of example to match released code
eranation Dec 20, 2014
a1a1361
SPARK-2641: Passing num executors to spark arguments from properties …
Dec 20, 2014
96d5b00
[SPARK-4140] Document dynamic allocation
Dec 20, 2014
4346a2b
[Minor] Build Failed: value defaultProperties not found
SaintBacchus Dec 20, 2014
665653d
[SPARK-2075][Core] Make the compiler generate same bytes code for Had…
zsxwing Dec 22, 2014
b896963
[SPARK-2075][Core] backport for branch-1.2
zsxwing Dec 22, 2014
31d42c4
[SPARK-4915][YARN] Fix classname to be specified for external shuffle…
Dec 22, 2014
70e69ef
[SPARK-4883][Shuffle] Add a name to the directoryCleaner thread
zsxwing Dec 22, 2014
c7396b5
[Minor] Improve some code in BroadcastTest for short
SaintBacchus Dec 22, 2014
4b2bded
[SPARK-4864] Add documentation to Netty-based configs
aarondav Dec 22, 2014
a8a8e0e
[SPARK-4920][UI]:current spark version in UI is not striking.
uncleGen Dec 22, 2014
58e3702
[SPARK-4818][Core] Add 'iterator' to reduce memory consumed by join
zsxwing Dec 22, 2014
f86fe08
[Docs] Minor typo fixes
nchammas Dec 23, 2014
9fb86b8
[SPARK-4931][Yarn][Docs] Fix the format of running-on-yarn.md
zsxwing Dec 23, 2014
ec11ffd
[SPARK-4834] [standalone] Clean up application files after app finishes.
Dec 23, 2014
e74ce14
[SPARK-4932] Add help comments in Analytics
maropu Dec 23, 2014
7b5ba85
[SPARK-4914][Build] Cleans lib_managed before compiling with Hive 0.13.1
liancheng Dec 23, 2014
6a46cc3
[SPARK-4730][YARN] Warn against deprecated YARN settings
Dec 23, 2014
01adf45
[SPARK-4802] [streaming] Remove receiverInfo once receiver is de-regi…
ilayaperumalg Dec 23, 2014
aa78c23
[SPARK-4671][Streaming]Do not replicate streaming block when WAL is e…
jerryshao Dec 23, 2014
1a4e2ba
[SPARK-4606] Send EOF to child JVM when there's no more data to read.
Dec 24, 2014
17d6f54
[SPARK-4873][Streaming] Use `Future.zip` instead of `Future.flatMap`(…
zsxwing Dec 25, 2014
475ab6e
Fix "Building Spark With Maven" link in README.md
dennyglee Dec 25, 2014
acf5c63
[SPARK-4537][Streaming] Expand StreamingSource to add more metrics
jerryshao Dec 26, 2014
391080b
SPARK-4971: Fix typo in BlockGenerator comment
CodingCat Dec 26, 2014
2e0af87
[SPARK-3787][BUILD] Assembly jar name is wrong when we build with sbt…
sarutak Dec 27, 2014
3c4acac
HOTFIX: Slight tweak on previous commit.
pwendell Dec 27, 2014
23d64cf
[SPARK-4952][Core]Handle ConcurrentModificationExceptions in SparkEnv…
witgo Dec 27, 2014
2cd446a
[SPARK-4966][YARN]The MemoryOverhead value is setted not correctly
XuTingjun Dec 29, 2014
7604666
[SPARK-4982][DOC] `spark.ui.retainedJobs` description is wrong in Spa…
wangxiaojing Dec 29, 2014
e81c869
SPARK-4968: takeOrdered to skip reduce step in case mappers return no…
Dec 29, 2014
e20d632
[SPARK-4920][UI] add version on master and worker page for standalone…
liyezhang556520 Dec 30, 2014
42809db
[SPARK-4882] Register PythonBroadcast with Kryo so that PySpark works…
JoshRosen Dec 30, 2014
cde8a31
[SPARK-4908][SQL] Prevent multiple concurrent hive native commands
marmbrus Dec 30, 2014
7a24541
[SPARK-4386] Improve performance when writing Parquet files
MickDavies Dec 30, 2014
edc96d8
[SPARK-4813][Streaming] Fix the issue that ContextWaiter didn't handl…
zsxwing Dec 30, 2014
ad3dc81
[SPARK-1010] Clean up uses of System.setProperty in unit tests
JoshRosen Dec 31, 2014
7c9c25b
[SPARK-4298][Core] - The spark-submit cannot read Main-Class from Man…
Dec 31, 2014
076de46
[HOTFIX] Disable Spark UI in SparkSubmitSuite tests
JoshRosen Dec 12, 2014
bd70ff9
[SPARK-4790][STREAMING] Fix ReceivedBlockTrackerSuite waits for old f…
harishreedharan Dec 31, 2014
14dbd83
[SPARK-5028][Streaming]Add total received and processed records metri…
jerryshao Dec 31, 2014
434ea00
[SPARK-5035] [Streaming] ReceiverMessage trait should extend Serializ…
JoshRosen Jan 1, 2015
da9a4b9
[HOTFIX] Bind web UI to ephemeral port in DriverSuite
JoshRosen Jan 1, 2015
33f0b14
Fixed typos in streaming-kafka-integration.md
Jan 2, 2015
93617dd
[SPARK-5058] Updated broken links
sigmoidanalytics Jan 4, 2015
9dbb62e
[SPARK-4787] Stop SparkContext if a DAGScheduler init error occurs
tigerquoll Jan 4, 2015
67e2eb6
[SPARK-4631] unit test for MQTT
Jan 5, 2015
a0bb88e
[SPARK-4835] Disable validateOutputSpecs for Spark Streaming jobs
JoshRosen Jan 5, 2015
f979205
[SPARK-4465] runAsSparkUser doesn't affect TaskRunner in Mesos enviro…
jongyoul Jan 5, 2015
cf55a2b
[SPARK-5089][PYSPARK][MLLIB] Fix vector convert
freeman-lab Jan 5, 2015
db83acb
[HOTFIX] Add missing SparkContext._ import to fix 1.2 build.
JoshRosen Jan 6, 2015
7a4be0b
[YARN][SPARK-4929] Bug fix: fix the yarn-client code to support HA
SaintBacchus Jan 7, 2015
1770c51
[SPARK-5132][Core]Correct stage Attempt Id key in stageInfofromJson
suyanNone Jan 7, 2015
755f9cc
[SPARK-5130][Deploy]Take yarn-cluster as cluster mode in spark-submit
WangTaoTheTonic Jan 8, 2015
71471bd
[SPARK-4973][CORE] Local directory in the driver of client-mode conti…
sarutak Jan 8, 2015
2f4e73d
SPARK-5136 [DOCS] Improve documentation around setting up Spark Intel…
srowen Jan 9, 2015
09eef3b
[SPARK-4925][SQL] Publish Spark SQL hive-thriftserver maven artifact
alexoss68 Jan 10, 2015
c6ea6d4
[SPARK-4943][SQL] Allow table name having dot for db/catalog
alexoss68 Jan 10, 2015
cce003d
[SPARK-5187][SQL] Fix caching of tables with HiveUDFs in the WHERE cl…
marmbrus Jan 10, 2015
f04ff9d
[SPARK-5181] do not print writing WAL log when WAL is disabled
CodingCat Jan 10, 2015
056149d
[SPARK-4951][Core] Fix the issue that a busy executor may be killed
zsxwing Jan 12, 2015
a1ee09e
[SPARK-5200] Disable web UI in Hive ThriftServer tests
JoshRosen Jan 12, 2015
558be07
[SPARK-5102][Core]subclass of MapStatus needs to be registered with Kryo
lianhuiwang Jan 12, 2015
5970f0b
[SPARK-5078] Optionally read from SPARK_LOCAL_HOSTNAME
marmbrus Jan 12, 2015
6d23af6
[SPARK-5049][SQL] Fix ordering of partition columns in ParquetTableScan
marmbrus Jan 12, 2015
7809683
[SPARK-5131][Streaming][DOC]: There is a discrepancy in WAL implement…
uncleGen Jan 13, 2015
1b6596e
[SPARK-5223] [MLlib] [PySpark] fix MapConverter and ListConverter in …
Jan 13, 2015
f7bbe29
[SPARK-5234][ml]examples for ml don't have sparkContext.stop
Jan 14, 2015
47fb0d0
[SPARK-5254][MLLIB] Update the user guide to position spark.ml better
mengxr Jan 15, 2015
3813547
[SPARK-5254][MLLIB] remove developers section from spark.ml guide
mengxr Jan 15, 2015
b3fe6df
[SPARK-5224] [PySpark] improve performance of parallelize list/ndarray
Jan 15, 2015
89a0990
[SPARK-4033][Examples]Input of the SparkPi too big causes the emptio…
SaintBacchus Jan 12, 2015
e38cb29
[SPARK-5201][CORE] deal with int overflow in the ParallelCollectionRD…
advancedxy Jan 16, 2015
473777e
[DOCS] Fix typo in return type of cogroup
srowen Jan 16, 2015
4a550ac
[SPARK-733] Add documentation on use of accumulators in lazy transfor…
Jan 16, 2015
94bafd8
[SPARK-5289]: Backport publishing of repl, yarn into branch-1.2.
pwendell Jan 18, 2015
5d5ee40
[SPARK-5282][mllib]: RowMatrix easily gets int overflow in the memory…
hhbyyh Jan 19, 2015
6599f50
[SPARK-4504][Examples] fix run-example failure if multiple assembly j…
gvramana Jan 19, 2015
228bf6c
[SPARK-4803] [streaming] Remove duplicate RegisterReceiver message
ilayaperumalg Jan 20, 2015
692dc5b
SPARK-4660: Use correct class loader in JavaSerializer (copy of PR #3…
jacek-lewandowski Jan 20, 2015
92c238c
[SPARK-4959][SQL] Attributes are case sensitive when using a select q…
chenghao-intel Jan 21, 2015
410b908
[SPARK-5275] [Streaming] include python source code
Jan 21, 2015
fd6266f
[HOTFIX] Update pom.xml to pull MapR's Hadoop version 2.4.1.
rkannan82 Jan 21, 2015
0c13eed
[SPARK-4759] Fix driver hanging from coalescing partitions
Dec 10, 2014
1d73017
[SPARK-4569] Rename 'externalSorting' in Aggregator
Dec 10, 2014
e90f6b5
[SPARK-4161]Spark shell class path is not correctly set if "spark.dri…
witgo Dec 10, 2014
37db20c
[SPARK-5064][GraphX] Add numEdges upperbound validation for R-MAT gra…
Jan 21, 2015
bb8bd11
[SPARK-5006][Deploy]spark.port.maxRetries doesn't work
WangTaoTheTonic Jan 13, 2015
079b3be
Make sure only owner can read / write to directories created for the …
Jan 21, 2015
dd18429
[SPARK-5355] make SparkConf thread-safe
Jan 22, 2015
cab410c
[SPARK-5147][Streaming] Delete the received data WAL log periodically
tdas Jan 22, 2015
5d07488
[HOTFIX] Fixed compilation error due to missing SparkContext._ implic…
tdas Jan 22, 2015
5aaf0e0
[SPARK-5233][Streaming] Fix error replaying of WAL introduced bug
jerryshao Jan 23, 2015
2ea782a
[SPARK-5063] More helpful error messages for several invalid operations
JoshRosen Jan 24, 2015
73cb806
[SPARK-5351][GraphX] Do not use Partitioner.defaultPartitioner as a p…
maropu Jan 24, 2015
ff2d7bd
[SPARK-5058] Part 2. Typos and broken URL
jongyoul Jan 24, 2015
1f8b718
[SPARK-5401] set executor ID before creating MetricsSystem
ryan-williams Jan 25, 2015
c573af4
[SPARK-5402] log executor ID at executor-construction time
ryan-williams Jan 25, 2015
a7e99ed
SPARK-3852 [DOCS] Document spark.driver.extra* configs
srowen Jan 25, 2015
7652809
SPARK-5382: Use SPARK_CONF_DIR in spark-class if it is defined
jacek-lewandowski Jan 25, 2015
e82e960
SPARK-5382: Use SPARK_CONF_DIR in spark-class and spark-submit, spark…
jacek-lewandowski Jan 25, 2015
f34c113
SPARK-4506 [DOCS] Addendum: Update more docs to reflect that standalo…
srowen Jan 25, 2015
8f55bee
[SPARK-5344][WebUI] HistoryServer cannot recognize that inprogress fi…
sarutak Jan 25, 2015
2a2da42
Revert "[SPARK-5344][WebUI] HistoryServer cannot recognize that inpro…
Jan 25, 2015
cf65620
SPARK-4430 [STREAMING] [TEST] Apache RAT Checks fail spuriously on te…
srowen Jan 26, 2015
ef6fe84
[SPARK-5355] use j.u.c.ConcurrentHashMap instead of TrieMap
Jan 26, 2015
b378e9a
SPARK-4147 [CORE] Reduce log4j dependency
srowen Jan 26, 2015
07c0fd1
Updating versions for Spark 1.2.1
pwendell Jan 27, 2015
e87eb2b
Preparing Spark release v1.2.1-rc1
Jan 27, 2015
adfed70
Preparing development version 1.2.2-SNAPSHOT
Jan 27, 2015
e8da342
Revert "Preparing development version 1.2.2-SNAPSHOT"
pwendell Jan 27, 2015
8c46100
Revert "Preparing Spark release v1.2.1-rc1"
pwendell Jan 27, 2015
3e2d7d3
Preparing Spark release v1.2.1-rc1
pwendell Jan 27, 2015
f53a431
Preparing development version 1.2.2-SNAPSHOT
pwendell Jan 27, 2015
8090448
SPARK-5308 [BUILD] MD5 / SHA1 hash format doesn't match standard Mave…
srowen Jan 27, 2015
fea9b43
[MLlib] fix python example of ALS in guide
Jan 27, 2015
063a4c5
Revert "Preparing development version 1.2.2-SNAPSHOT"
pwendell Jan 28, 2015
4026bba
Revert "Preparing Spark release v1.2.1-rc1"
pwendell Jan 28, 2015
b77f876
Preparing Spark release v1.2.1-rc2
pwendell Jan 28, 2015
0a16aba
Preparing development version 1.2.2-SNAPSHOT
pwendell Jan 28, 2015
00746a5
[Docs] Fix Building Spark link text
nchammas Feb 2, 2015
b978c9f
Disabling Utils.chmod700 for Windows
MartinWeindel Feb 2, 2015
5486440
[SPARK-5195][sql]Update HiveMetastoreCatalog.scala(override the Metas…
seayi Feb 3, 2015
88e0f2d
Revert "[SPARK-5195][sql]Update HiveMetastoreCatalog.scala(override t…
pwendell Feb 3, 2015
d944c0b
Revert "Preparing development version 1.2.2-SNAPSHOT"
pwendell Feb 3, 2015
a64c7a8
Revert "Preparing Spark release v1.2.1-rc2"
pwendell Feb 3, 2015
b6eaf77
Preparing Spark release v1.2.1-rc3
pwendell Feb 3, 2015
f15b736
[SPARK-5691] Fixing wrong data structure lookup for dupe app registra…
mccheah Feb 9, 2015
b616336
[SPARK-4808] Removing minimum number of elements read before spill check
mccheah Feb 20, 2015
c060dca
Preparing Palantir Spark release v1.2.1-palantir1
mingyukim Mar 9, 2015
9f24d73
Preparing Palantir Spark release v1.2.1-palantir2
mingyukim Mar 24, 2015
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
5 changes: 4 additions & 1 deletion .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@
*.ipr
*.iml
*.iws
*.pyc
.idea/
.idea_modules/
sbt/*.jar
Expand Down Expand Up @@ -49,7 +50,9 @@ dependency-reduced-pom.xml
checkpoint
derby.log
dist/
spark-*-bin.tar.gz
dev/create-release/*txt
dev/create-release/*final
spark-*-bin-*.tgz
unit-tests.log
/lib/
rat-results.txt
Expand Down
2 changes: 2 additions & 0 deletions .rat-excludes
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,7 @@ SparkImports.scala
SparkJLineCompletion.scala
SparkJLineReader.scala
SparkMemberHandlers.scala
SparkReplReporter.scala
sbt
sbt-launch-lib.bash
plugins.sbt
Expand All @@ -63,3 +64,4 @@ dist/*
logs
.*scalastyle-output.xml
.*dependency-reduced-pom.xml
known_translations
24 changes: 22 additions & 2 deletions LICENSE
Original file line number Diff line number Diff line change
Expand Up @@ -646,7 +646,8 @@ THE SOFTWARE.

========================================================================
For Scala Interpreter classes (all .scala files in repl/src/main/scala
except for Main.Scala, SparkHelper.scala and ExecutorClassLoader.scala):
except for Main.Scala, SparkHelper.scala and ExecutorClassLoader.scala),
and for SerializableMapWrapper in JavaUtils.scala:
========================================================================

Copyright (c) 2002-2013 EPFL
Expand Down Expand Up @@ -754,7 +755,7 @@ SUCH DAMAGE.


========================================================================
For Timsort (core/src/main/java/org/apache/spark/util/collection/Sorter.java):
For Timsort (core/src/main/java/org/apache/spark/util/collection/TimSort.java):
========================================================================
Copyright (C) 2008 The Android Open Source Project

Expand All @@ -771,6 +772,25 @@ See the License for the specific language governing permissions and
limitations under the License.


========================================================================
For LimitedInputStream
(network/common/src/main/java/org/apache/spark/network/util/LimitedInputStream.java):
========================================================================
Copyright (C) 2007 The Guava Authors

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.


========================================================================
BSD-style licenses
========================================================================
Expand Down
5 changes: 3 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,8 @@ and Spark Streaming for stream processing.
## Online Documentation

You can find the latest Spark documentation, including a programming
guide, on the [project web page](http://spark.apache.org/documentation.html).
guide, on the [project web page](http://spark.apache.org/documentation.html)
and [project wiki](https://cwiki.apache.org/confluence/display/SPARK).
This README file only contains basic setup instructions.

## Building Spark
Expand All @@ -25,7 +26,7 @@ To build Spark and its example programs, run:

(You do not need to do this if you downloaded a pre-built package.)
More detailed documentation is available from the project site, at
["Building Spark with Maven"](http://spark.apache.org/docs/latest/building-with-maven.html).
["Building Spark"](http://spark.apache.org/docs/latest/building-spark.html).

## Interactive Scala Shell

Expand Down
15 changes: 10 additions & 5 deletions assembly/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@
<parent>
<groupId>org.apache.spark</groupId>
<artifactId>spark-parent</artifactId>
<version>1.2.0-SNAPSHOT</version>
<version>1.2.1-palantir2</version>
<relativePath>../pom.xml</relativePath>
</parent>

Expand Down Expand Up @@ -66,22 +66,22 @@
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-repl_${scala.binary.version}</artifactId>
<artifactId>spark-streaming_${scala.binary.version}</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-streaming_${scala.binary.version}</artifactId>
<artifactId>spark-graphx_${scala.binary.version}</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-graphx_${scala.binary.version}</artifactId>
<artifactId>spark-sql_${scala.binary.version}</artifactId>
<version>${project.version}</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_${scala.binary.version}</artifactId>
<artifactId>spark-repl_${scala.binary.version}</artifactId>
<version>${project.version}</version>
</dependency>
</dependencies>
Expand Down Expand Up @@ -197,6 +197,11 @@
<artifactId>spark-hive_${scala.binary.version}</artifactId>
<version>${project.version}</version>
</dependency>
</dependencies>
</profile>
<profile>
<id>hive-thriftserver</id>
<dependencies>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-hive-thriftserver_${scala.binary.version}</artifactId>
Expand Down
2 changes: 1 addition & 1 deletion bagel/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@
<parent>
<groupId>org.apache.spark</groupId>
<artifactId>spark-parent</artifactId>
<version>1.2.0-SNAPSHOT</version>
<version>1.2.1-palantir2</version>
<relativePath>../pom.xml</relativePath>
</parent>

Expand Down
21 changes: 21 additions & 0 deletions bin/beeline.cmd
Original file line number Diff line number Diff line change
@@ -0,0 +1,21 @@
@echo off

rem
rem Licensed to the Apache Software Foundation (ASF) under one or more
rem contributor license agreements. See the NOTICE file distributed with
rem this work for additional information regarding copyright ownership.
rem The ASF licenses this file to You under the Apache License, Version 2.0
rem (the "License"); you may not use this file except in compliance with
rem the License. You may obtain a copy of the License at
rem
rem http://www.apache.org/licenses/LICENSE-2.0
rem
rem Unless required by applicable law or agreed to in writing, software
rem distributed under the License is distributed on an "AS IS" BASIS,
rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
rem See the License for the specific language governing permissions and
rem limitations under the License.
rem

set SPARK_HOME=%~dp0..
cmd /V /E /C %SPARK_HOME%\bin\spark-class.cmd org.apache.hive.beeline.BeeLine %*
73 changes: 37 additions & 36 deletions bin/compute-classpath.sh
Original file line number Diff line number Diff line change
Expand Up @@ -20,8 +20,6 @@
# This script computes Spark's classpath and prints it to stdout; it's used by both the "run"
# script and the ExecutorRunner in standalone cluster mode.

SCALA_VERSION=2.10

# Figure out where Spark is installed
FWDIR="$(cd "`dirname "$0"`"/..; pwd)"

Expand All @@ -36,7 +34,7 @@ else
CLASSPATH="$CLASSPATH:$FWDIR/conf"
fi

ASSEMBLY_DIR="$FWDIR/assembly/target/scala-$SCALA_VERSION"
ASSEMBLY_DIR="$FWDIR/assembly/target/scala-$SPARK_SCALA_VERSION"

if [ -n "$JAVA_HOME" ]; then
JAR_CMD="$JAVA_HOME/bin/jar"
Expand All @@ -48,19 +46,19 @@ fi
if [ -n "$SPARK_PREPEND_CLASSES" ]; then
echo "NOTE: SPARK_PREPEND_CLASSES is set, placing locally compiled Spark"\
"classes ahead of assembly." >&2
CLASSPATH="$CLASSPATH:$FWDIR/core/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/core/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/core/target/jars/*"
CLASSPATH="$CLASSPATH:$FWDIR/repl/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/mllib/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/bagel/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/graphx/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/streaming/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/tools/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/catalyst/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/core/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/hive/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/hive-thriftserver/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/yarn/stable/target/scala-$SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/repl/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/mllib/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/bagel/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/graphx/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/streaming/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/tools/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/catalyst/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/core/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/hive/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/hive-thriftserver/target/scala-$SPARK_SCALA_VERSION/classes"
CLASSPATH="$CLASSPATH:$FWDIR/yarn/stable/target/scala-$SPARK_SCALA_VERSION/classes"
fi

# Use spark-assembly jar from either RELEASE or assembly directory
Expand All @@ -70,22 +68,25 @@ else
assembly_folder="$ASSEMBLY_DIR"
fi

num_jars="$(ls "$assembly_folder" | grep "spark-assembly.*hadoop.*\.jar" | wc -l)"
if [ "$num_jars" -eq "0" ]; then
echo "Failed to find Spark assembly in $assembly_folder"
echo "You need to build Spark before running this program."
exit 1
fi
num_jars=0

for f in ${assembly_folder}/spark-assembly*hadoop*.jar; do
if [[ ! -e "$f" ]]; then
echo "Failed to find Spark assembly in $assembly_folder" 1>&2
echo "You need to build Spark before running this program." 1>&2
exit 1
fi
ASSEMBLY_JAR="$f"
num_jars=$((num_jars+1))
done

if [ "$num_jars" -gt "1" ]; then
jars_list=$(ls "$assembly_folder" | grep "spark-assembly.*hadoop.*.jar")
echo "Found multiple Spark assembly jars in $assembly_folder:"
echo "$jars_list"
echo "Please remove all but one jar."
echo "Found multiple Spark assembly jars in $assembly_folder:" 1>&2
ls ${assembly_folder}/spark-assembly*hadoop*.jar 1>&2
echo "Please remove all but one jar." 1>&2
exit 1
fi

ASSEMBLY_JAR="$(ls "$assembly_folder"/spark-assembly*hadoop*.jar 2>/dev/null)"

# Verify that versions of java used to build the jars and run Spark are compatible
jar_error_check=$("$JAR_CMD" -tf "$ASSEMBLY_JAR" nonexistent/class/path 2>&1)
if [[ "$jar_error_check" =~ "invalid CEN header" ]]; then
Expand Down Expand Up @@ -123,15 +124,15 @@ fi

# Add test classes if we're running from SBT or Maven with SPARK_TESTING set to 1
if [[ $SPARK_TESTING == 1 ]]; then
CLASSPATH="$CLASSPATH:$FWDIR/core/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/repl/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/mllib/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/bagel/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/graphx/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/streaming/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/catalyst/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/core/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/hive/target/scala-$SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/core/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/repl/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/mllib/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/bagel/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/graphx/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/streaming/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/catalyst/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/core/target/scala-$SPARK_SCALA_VERSION/test-classes"
CLASSPATH="$CLASSPATH:$FWDIR/sql/hive/target/scala-$SPARK_SCALA_VERSION/test-classes"
fi

# Add hadoop conf dir if given -- otherwise FileSystem.*, etc fail !
Expand Down
20 changes: 20 additions & 0 deletions bin/load-spark-env.sh
Original file line number Diff line number Diff line change
Expand Up @@ -36,3 +36,23 @@ if [ -z "$SPARK_ENV_LOADED" ]; then
set +a
fi
fi

# Setting SPARK_SCALA_VERSION if not already set.

if [ -z "$SPARK_SCALA_VERSION" ]; then

ASSEMBLY_DIR2="$FWDIR/assembly/target/scala-2.11"
ASSEMBLY_DIR1="$FWDIR/assembly/target/scala-2.10"

if [[ -d "$ASSEMBLY_DIR2" && -d "$ASSEMBLY_DIR1" ]]; then
echo -e "Presence of build for both scala versions(SCALA 2.10 and SCALA 2.11) detected." 1>&2
echo -e 'Either clean one of them or, export SPARK_SCALA_VERSION=2.11 in spark-env.sh.' 1>&2
exit 1
fi

if [ -d "$ASSEMBLY_DIR2" ]; then
export SPARK_SCALA_VERSION="2.11"
else
export SPARK_SCALA_VERSION="2.10"
fi
fi
8 changes: 2 additions & 6 deletions bin/pyspark
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ export SPARK_HOME="$FWDIR"

source "$FWDIR/bin/utils.sh"

SCALA_VERSION=2.10
source "$FWDIR"/bin/load-spark-env.sh

function usage() {
echo "Usage: ./bin/pyspark [options]" 1>&2
Expand All @@ -40,16 +40,14 @@ fi
# Exit if the user hasn't compiled Spark
if [ ! -f "$FWDIR/RELEASE" ]; then
# Exit if the user hasn't compiled Spark
ls "$FWDIR"/assembly/target/scala-$SCALA_VERSION/spark-assembly*hadoop*.jar >& /dev/null
ls "$FWDIR"/assembly/target/scala-$SPARK_SCALA_VERSION/spark-assembly*hadoop*.jar >& /dev/null
if [[ $? != 0 ]]; then
echo "Failed to find Spark assembly in $FWDIR/assembly/target" 1>&2
echo "You need to build Spark before running this program" 1>&2
exit 1
fi
fi

. "$FWDIR"/bin/load-spark-env.sh

# In Spark <= 1.1, setting IPYTHON=1 would cause the driver to be launched using the `ipython`
# executable, while the worker would still be launched using PYSPARK_PYTHON.
#
Expand Down Expand Up @@ -134,7 +132,5 @@ if [[ "$1" =~ \.py$ ]]; then
gatherSparkSubmitOpts "$@"
exec "$FWDIR"/bin/spark-submit "${SUBMISSION_OPTS[@]}" "$primary" "${APPLICATION_OPTS[@]}"
else
# PySpark shell requires special handling downstream
export PYSPARK_SHELL=1
exec "$PYSPARK_DRIVER_PYTHON" $PYSPARK_DRIVER_PYTHON_OPTS
fi
1 change: 0 additions & 1 deletion bin/pyspark2.cmd
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,6 @@ for /f %%i in ('echo %1^| findstr /R "\.py"') do (
)

if [%PYTHON_FILE%] == [] (
set PYSPARK_SHELL=1
if [%IPYTHON%] == [1] (
ipython %IPYTHON_OPTS%
) else (
Expand Down
Loading