|
1 |
| -@echo off |
2 |
| - |
3 |
| -rem |
4 |
| -rem Licensed to the Apache Software Foundation (ASF) under one or more |
5 |
| -rem contributor license agreements. See the NOTICE file distributed with |
6 |
| -rem this work for additional information regarding copyright ownership. |
7 |
| -rem The ASF licenses this file to You under the Apache License, Version 2.0 |
8 |
| -rem (the "License"); you may not use this file except in compliance with |
9 |
| -rem the License. You may obtain a copy of the License at |
10 |
| -rem |
11 |
| -rem http://www.apache.org/licenses/LICENSE-2.0 |
12 |
| -rem |
13 |
| -rem Unless required by applicable law or agreed to in writing, software |
14 |
| -rem distributed under the License is distributed on an "AS IS" BASIS, |
15 |
| -rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
16 |
| -rem See the License for the specific language governing permissions and |
17 |
| -rem limitations under the License. |
18 |
| -rem |
19 |
| - |
20 |
| -rem This script computes Spark's classpath and prints it to stdout; it's used by both the "run" |
21 |
| -rem script and the ExecutorRunner in standalone cluster mode. |
22 |
| - |
23 |
| -rem If we're called from spark-class2.cmd, it already set enabledelayedexpansion and setting |
24 |
| -rem it here would stop us from affecting its copy of the CLASSPATH variable; otherwise we |
25 |
| -rem need to set it here because we use !datanucleus_jars! below. |
26 |
| -if "%DONT_PRINT_CLASSPATH%"=="1" goto skip_delayed_expansion |
27 |
| -setlocal enabledelayedexpansion |
28 |
| -:skip_delayed_expansion |
29 |
| - |
30 |
| -set SCALA_VERSION=2.10 |
31 |
| - |
32 |
| -rem Figure out where the Spark framework is installed |
33 |
| -set FWDIR=%~dp0..\ |
34 |
| - |
35 |
| -rem Load environment variables from conf\spark-env.cmd, if it exists |
36 |
| -if exist "%FWDIR%conf\spark-env.cmd" call "%FWDIR%conf\spark-env.cmd" |
37 |
| - |
38 |
| -rem Build up classpath |
39 |
| -set CLASSPATH=%SPARK_CLASSPATH%;%SPARK_SUBMIT_CLASSPATH% |
40 |
| - |
41 |
| -if not "x%SPARK_CONF_DIR%"=="x" ( |
42 |
| - set CLASSPATH=%CLASSPATH%;%SPARK_CONF_DIR% |
43 |
| -) else ( |
44 |
| - set CLASSPATH=%CLASSPATH%;%FWDIR%conf |
45 |
| -) |
46 |
| - |
47 |
| -if exist "%FWDIR%RELEASE" ( |
48 |
| - for %%d in ("%FWDIR%lib\spark-assembly*.jar") do ( |
49 |
| - set ASSEMBLY_JAR=%%d |
50 |
| - ) |
51 |
| -) else ( |
52 |
| - for %%d in ("%FWDIR%assembly\target\scala-%SCALA_VERSION%\spark-assembly*hadoop*.jar") do ( |
53 |
| - set ASSEMBLY_JAR=%%d |
54 |
| - ) |
55 |
| -) |
56 |
| - |
57 |
| -set CLASSPATH=%CLASSPATH%;%ASSEMBLY_JAR% |
58 |
| - |
59 |
| -rem When Hive support is needed, Datanucleus jars must be included on the classpath. |
60 |
| -rem Datanucleus jars do not work if only included in the uber jar as plugin.xml metadata is lost. |
61 |
| -rem Both sbt and maven will populate "lib_managed/jars/" with the datanucleus jars when Spark is |
62 |
| -rem built with Hive, so look for them there. |
63 |
| -if exist "%FWDIR%RELEASE" ( |
64 |
| - set datanucleus_dir=%FWDIR%lib |
65 |
| -) else ( |
66 |
| - set datanucleus_dir=%FWDIR%lib_managed\jars |
67 |
| -) |
68 |
| -set "datanucleus_jars=" |
69 |
| -for %%d in ("%datanucleus_dir%\datanucleus-*.jar") do ( |
70 |
| - set datanucleus_jars=!datanucleus_jars!;%%d |
71 |
| -) |
72 |
| -set CLASSPATH=%CLASSPATH%;%datanucleus_jars% |
73 |
| - |
74 |
| -set SPARK_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\classes |
75 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\classes |
76 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\classes |
77 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\classes |
78 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\classes |
79 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\classes |
80 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%tools\target\scala-%SCALA_VERSION%\classes |
81 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\classes |
82 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\classes |
83 |
| -set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\classes |
84 |
| - |
85 |
| -set SPARK_TEST_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\test-classes |
86 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\test-classes |
87 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\test-classes |
88 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\test-classes |
89 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\test-classes |
90 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\test-classes |
91 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\test-classes |
92 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\test-classes |
93 |
| -set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\test-classes |
94 |
| - |
95 |
| -if "x%SPARK_TESTING%"=="x1" ( |
96 |
| - rem Add test clases to path - note, add SPARK_CLASSES and SPARK_TEST_CLASSES before CLASSPATH |
97 |
| - rem so that local compilation takes precedence over assembled jar |
98 |
| - set CLASSPATH=%SPARK_CLASSES%;%SPARK_TEST_CLASSES%;%CLASSPATH% |
99 |
| -) |
100 |
| - |
101 |
| -rem Add hadoop conf dir - else FileSystem.*, etc fail |
102 |
| -rem Note, this assumes that there is either a HADOOP_CONF_DIR or YARN_CONF_DIR which hosts |
103 |
| -rem the configurtion files. |
104 |
| -if "x%HADOOP_CONF_DIR%"=="x" goto no_hadoop_conf_dir |
105 |
| - set CLASSPATH=%CLASSPATH%;%HADOOP_CONF_DIR% |
106 |
| -:no_hadoop_conf_dir |
107 |
| - |
108 |
| -if "x%YARN_CONF_DIR%"=="x" goto no_yarn_conf_dir |
109 |
| - set CLASSPATH=%CLASSPATH%;%YARN_CONF_DIR% |
110 |
| -:no_yarn_conf_dir |
111 |
| - |
112 |
| -rem A bit of a hack to allow calling this script within run2.cmd without seeing output |
113 |
| -if "%DONT_PRINT_CLASSPATH%"=="1" goto exit |
114 |
| - |
115 |
| -echo %CLASSPATH% |
116 |
| - |
117 |
| -:exit |
| 1 | +@echo off |
| 2 | + |
| 3 | +rem |
| 4 | +rem Licensed to the Apache Software Foundation (ASF) under one or more |
| 5 | +rem contributor license agreements. See the NOTICE file distributed with |
| 6 | +rem this work for additional information regarding copyright ownership. |
| 7 | +rem The ASF licenses this file to You under the Apache License, Version 2.0 |
| 8 | +rem (the "License"); you may not use this file except in compliance with |
| 9 | +rem the License. You may obtain a copy of the License at |
| 10 | +rem |
| 11 | +rem http://www.apache.org/licenses/LICENSE-2.0 |
| 12 | +rem |
| 13 | +rem Unless required by applicable law or agreed to in writing, software |
| 14 | +rem distributed under the License is distributed on an "AS IS" BASIS, |
| 15 | +rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 16 | +rem See the License for the specific language governing permissions and |
| 17 | +rem limitations under the License. |
| 18 | +rem |
| 19 | + |
| 20 | +rem This script computes Spark's classpath and prints it to stdout; it's used by both the "run" |
| 21 | +rem script and the ExecutorRunner in standalone cluster mode. |
| 22 | + |
| 23 | +rem If we're called from spark-class2.cmd, it already set enabledelayedexpansion and setting |
| 24 | +rem it here would stop us from affecting its copy of the CLASSPATH variable; otherwise we |
| 25 | +rem need to set it here because we use !datanucleus_jars! below. |
| 26 | +if "%DONT_PRINT_CLASSPATH%"=="1" goto skip_delayed_expansion |
| 27 | +setlocal enabledelayedexpansion |
| 28 | +:skip_delayed_expansion |
| 29 | + |
| 30 | +set SCALA_VERSION=2.10 |
| 31 | + |
| 32 | +rem Figure out where the Spark framework is installed |
| 33 | +set FWDIR=%~dp0..\ |
| 34 | + |
| 35 | +rem Load environment variables from conf\spark-env.cmd, if it exists |
| 36 | +if exist "%FWDIR%conf\spark-env.cmd" call "%FWDIR%conf\spark-env.cmd" |
| 37 | + |
| 38 | +rem Build up classpath |
| 39 | +set CLASSPATH=%SPARK_CLASSPATH%;%SPARK_SUBMIT_CLASSPATH% |
| 40 | + |
| 41 | +if not "x%SPARK_CONF_DIR%"=="x" ( |
| 42 | + set CLASSPATH=%CLASSPATH%;%SPARK_CONF_DIR% |
| 43 | +) else ( |
| 44 | + set CLASSPATH=%CLASSPATH%;%FWDIR%conf |
| 45 | +) |
| 46 | + |
| 47 | +if exist "%FWDIR%RELEASE" ( |
| 48 | + for %%d in ("%FWDIR%lib\spark-assembly*.jar") do ( |
| 49 | + set ASSEMBLY_JAR=%%d |
| 50 | + ) |
| 51 | +) else ( |
| 52 | + for %%d in ("%FWDIR%assembly\target\scala-%SCALA_VERSION%\spark-assembly*hadoop*.jar") do ( |
| 53 | + set ASSEMBLY_JAR=%%d |
| 54 | + ) |
| 55 | +) |
| 56 | + |
| 57 | +set CLASSPATH=%CLASSPATH%;%ASSEMBLY_JAR% |
| 58 | + |
| 59 | +rem When Hive support is needed, Datanucleus jars must be included on the classpath. |
| 60 | +rem Datanucleus jars do not work if only included in the uber jar as plugin.xml metadata is lost. |
| 61 | +rem Both sbt and maven will populate "lib_managed/jars/" with the datanucleus jars when Spark is |
| 62 | +rem built with Hive, so look for them there. |
| 63 | +if exist "%FWDIR%RELEASE" ( |
| 64 | + set datanucleus_dir=%FWDIR%lib |
| 65 | +) else ( |
| 66 | + set datanucleus_dir=%FWDIR%lib_managed\jars |
| 67 | +) |
| 68 | +set "datanucleus_jars=" |
| 69 | +for %%d in ("%datanucleus_dir%\datanucleus-*.jar") do ( |
| 70 | + set datanucleus_jars=!datanucleus_jars!;%%d |
| 71 | +) |
| 72 | +set CLASSPATH=%CLASSPATH%;%datanucleus_jars% |
| 73 | + |
| 74 | +set SPARK_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\classes |
| 75 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\classes |
| 76 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\classes |
| 77 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\classes |
| 78 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\classes |
| 79 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\classes |
| 80 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%tools\target\scala-%SCALA_VERSION%\classes |
| 81 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\classes |
| 82 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\classes |
| 83 | +set SPARK_CLASSES=%SPARK_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\classes |
| 84 | + |
| 85 | +set SPARK_TEST_CLASSES=%FWDIR%core\target\scala-%SCALA_VERSION%\test-classes |
| 86 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%repl\target\scala-%SCALA_VERSION%\test-classes |
| 87 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%mllib\target\scala-%SCALA_VERSION%\test-classes |
| 88 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%bagel\target\scala-%SCALA_VERSION%\test-classes |
| 89 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%graphx\target\scala-%SCALA_VERSION%\test-classes |
| 90 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%streaming\target\scala-%SCALA_VERSION%\test-classes |
| 91 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\catalyst\target\scala-%SCALA_VERSION%\test-classes |
| 92 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\core\target\scala-%SCALA_VERSION%\test-classes |
| 93 | +set SPARK_TEST_CLASSES=%SPARK_TEST_CLASSES%;%FWDIR%sql\hive\target\scala-%SCALA_VERSION%\test-classes |
| 94 | + |
| 95 | +if "x%SPARK_TESTING%"=="x1" ( |
| 96 | + rem Add test clases to path - note, add SPARK_CLASSES and SPARK_TEST_CLASSES before CLASSPATH |
| 97 | + rem so that local compilation takes precedence over assembled jar |
| 98 | + set CLASSPATH=%SPARK_CLASSES%;%SPARK_TEST_CLASSES%;%CLASSPATH% |
| 99 | +) |
| 100 | + |
| 101 | +rem Add hadoop conf dir - else FileSystem.*, etc fail |
| 102 | +rem Note, this assumes that there is either a HADOOP_CONF_DIR or YARN_CONF_DIR which hosts |
| 103 | +rem the configurtion files. |
| 104 | +if "x%HADOOP_CONF_DIR%"=="x" goto no_hadoop_conf_dir |
| 105 | + set CLASSPATH=%CLASSPATH%;%HADOOP_CONF_DIR% |
| 106 | +:no_hadoop_conf_dir |
| 107 | + |
| 108 | +if "x%YARN_CONF_DIR%"=="x" goto no_yarn_conf_dir |
| 109 | + set CLASSPATH=%CLASSPATH%;%YARN_CONF_DIR% |
| 110 | +:no_yarn_conf_dir |
| 111 | + |
| 112 | +rem A bit of a hack to allow calling this script within run2.cmd without seeing output |
| 113 | +if "%DONT_PRINT_CLASSPATH%"=="1" goto exit |
| 114 | + |
| 115 | +echo %CLASSPATH% |
| 116 | + |
| 117 | +:exit |
0 commit comments