Skip to content

Commit

Permalink
Revert "[SPARK-1150] fix repo location in create script"
Browse files Browse the repository at this point in the history
This reverts commit 9aa0957.
  • Loading branch information
pwendell committed Mar 2, 2014
1 parent 9aa0957 commit ec992e1
Show file tree
Hide file tree
Showing 3 changed files with 5 additions and 11 deletions.
2 changes: 1 addition & 1 deletion conf/spark-env.sh.template
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
# - SPARK_MASTER_IP, to bind the master to a different IP address or hostname
# - SPARK_MASTER_PORT / SPARK_MASTER_WEBUI_PORT, to use non-default ports
# - SPARK_WORKER_CORES, to set the number of cores to use on this machine
# - SPARK_WORKER_MEM, to set how much memory to use (e.g. 1000m, 2g)
# - SPARK_WORKER_MEMORY, to set how much memory to use (e.g. 1000m, 2g)
# - SPARK_WORKER_PORT / SPARK_WORKER_WEBUI_PORT
# - SPARK_WORKER_INSTANCES, to set the number of worker processes per node
# - SPARK_WORKER_DIR, to set the working directory of worker processes
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,15 +18,13 @@
package org.apache.spark.deploy.worker

import java.lang.management.ManagementFactory
import org.apache.spark.Logging

import org.apache.spark.util.{IntParam, MemoryParam, Utils}

/**
* Command-line parser for the master.
*/
private[spark] class WorkerArguments(args: Array[String]) extends Logging {
initLogging()
private[spark] class WorkerArguments(args: Array[String]) {
var host = Utils.localHostName()
var port = 0
var webUiPort = 8081
Expand All @@ -42,13 +40,9 @@ private[spark] class WorkerArguments(args: Array[String]) extends Logging {
if (System.getenv("SPARK_WORKER_CORES") != null) {
cores = System.getenv("SPARK_WORKER_CORES").toInt
}
if (System.getenv("SPARK_WORKER_MEM") != null) {
memory = Utils.memoryStringToMb(System.getenv("SPARK_WORKER_MEM"))
} else if (System.getenv("SPARK_WORKER_MEMORY") != null) {
logWarning("SPARK_WORKER_MEMORY is deprecated. Please use SPARK_WORKER_MEM instead")
if (System.getenv("SPARK_WORKER_MEMORY") != null) {
memory = Utils.memoryStringToMb(System.getenv("SPARK_WORKER_MEMORY"))
}

if (System.getenv("SPARK_WORKER_WEBUI_PORT") != null) {
webUiPort = System.getenv("SPARK_WORKER_WEBUI_PORT").toInt
}
Expand Down
4 changes: 2 additions & 2 deletions docs/spark-standalone.md
Original file line number Diff line number Diff line change
Expand Up @@ -104,8 +104,8 @@ You can optionally configure the cluster further by setting environment variable
<td>Total number of cores to allow Spark applications to use on the machine (default: all available cores).</td>
</tr>
<tr>
<td><code>SPARK_WORKER_MEM</code></td>
<td>Total amount of memory to allow Spark applications to use on the machine, e.g. <code>1000m</code>, <code>2g</code> (default: total memory minus 1 GB); note that each application's <i>individual</i> memory is configured using its <code>spark.executor.memory</code> property. The old variable </code>SPARK_WORKER_MEMORY</code> has been deprecated.</td>
<td><code>SPARK_WORKER_MEMORY</code></td>
<td>Total amount of memory to allow Spark applications to use on the machine, e.g. <code>1000m</code>, <code>2g</code> (default: total memory minus 1 GB); note that each application's <i>individual</i> memory is configured using its <code>spark.executor.memory</code> property.</td>
</tr>
<tr>
<td><code>SPARK_WORKER_WEBUI_PORT</code></td>
Expand Down

0 comments on commit ec992e1

Please sign in to comment.