@@ -13,7 +13,7 @@ def get_run_sh(hdfs_env, hdfs_input, path_src, time_str, app_name, mem):
13
13
14
14
# python_cmd = 'python -u main.py {} {} | tee {}/log.txt'.format(path_input, path_output, path_output)
15
15
# python_cmd = 'python -u fengdian_ensemble.py {} -1 {} | tee {}/log.txt'.format(path_output, path_input, path_output)
16
- python_cmd = f'python -u run.py { path_input } | tee { path_output } /kaggle_house_price.log'
16
+ python_cmd = f'python -u run.py { path_input } { path_output } | tee { path_output } /kaggle_house_price.log'
17
17
18
18
cmd_hdfs_get = 'hdfs dfs -get {}/{}/ .' .format (hdfs_input , name_output )
19
19
file = open ('cmd_hdfs_get.sh' , 'a' )
@@ -63,7 +63,7 @@ def get_run_sh(hdfs_env, hdfs_input, path_src, time_str, app_name, mem):
63
63
s += ['echo "[+] output......"' ]
64
64
s += ['mkdir {}' .format (name_output )]
65
65
s += ['hdfs dfs -put -f -p {}/ {}/' .format (name_output , hdfs_input )]
66
- s += ['hdfs dfs -put -f src .tar.gz {}/{}/' .format (hdfs_input , name_output )]
66
+ s += ['hdfs dfs -put -f autox .tar.gz {}/{}/' .format (hdfs_input , name_output )]
67
67
s += ['hdfs dfs -put -f run.sh {}/{}/' .format (hdfs_input , name_output )]
68
68
s += ['hdfs dfs -put -f yarn.sh {}/{}/' .format (hdfs_input , name_output )]
69
69
s += ['hdfs dfs -put -f submit.py {}/{}/' .format (hdfs_input , name_output )]
@@ -101,7 +101,7 @@ def get_yarn_sh(appname, mem=65536, queue='pico'):
101
101
s += [' --num_containers=1 \\ ' ]
102
102
s += [' --shell_env HADOOP_USER_NAME=`whoami`\\ ' ]
103
103
s += [' --shell_env WEBHDFS_USER=`whoami` \\ ' ]
104
- s += [' --file src .tar.gz \\ ' ]
104
+ s += [' --file autox .tar.gz \\ ' ]
105
105
s += [' --file submit.py \\ ' ]
106
106
s += [' --file run.sh \\ ' ]
107
107
s += [' --file yarn.sh' ]
@@ -159,7 +159,7 @@ def tar_src(path_src):
159
159
return path_src
160
160
161
161
162
- path_src = 'src '
162
+ path_src = 'autox '
163
163
path_src = tar_src (path_src )
164
164
if path_src is None :
165
165
print ('[+] src not exist!' )
0 commit comments