= RpcEnvConfig :page-toclevels: -1
[[creating-instance]] RpcEnvConfig is a configuration of an rpc:RpcEnv.md:
- [[conf]] SparkConf.md
- [[name]] System Name
- [[bindAddress]] Bind Address
- [[advertiseAddress]] Advertised Address
- [[port]] Port
- [[securityManager]] SecurityManager
- [[numUsableCores]] Number of CPU cores
RpcEnvConfig is created when RpcEnv utility is used to rpc:RpcEnv.md#create[create an RpcEnv] (using rpc:RpcEnvFactory.md).
== [[clientMode]] Client Mode
When an RPC Environment is initialized core:SparkEnv.md#createDriverEnv[as part of the initialization of the driver] or core:SparkEnv.md#createExecutorEnv[executors] (using
false for the driver and
true for executors.
Copied (almost verbatim) from https://issues.apache.org/jira/browse/SPARK-10997[SPARK-10997 Netty-based RPC env should support a "client-only" mode] and the https://github.com/apache/spark/commit/71d1c907dec446db566b19f912159fd8f46deb7d[commit]:
"Client mode" means the RPC env will not listen for incoming connections.
This allows certain processes in the Spark stack (such as Executors or tha YARN client-mode AM) to act as pure clients when using the netty-based RPC backend, reducing the number of sockets Spark apps need to use and also the number of open ports.
The AM connects to the driver in "client mode", and that connection is used for all driver -- AM communication, and so the AM is properly notified when the connection goes down.
In "general", non-YARN case,
clientMode flag is therefore enabled for executors and disabled for the driver.
In Spark on YARN in
client deploy mode,
clientMode flag is however enabled explicitly when Spark on YARN's spark-yarn-applicationmaster.md#runExecutorLauncher-sparkYarnAM[ApplicationMaster] creates the
sparkYarnAM RPC Environment.