I have a number of hive queries that my system executes on a regular basis. When you look at the job tracker, they show up as "SELECT field, other_field ..... (Stage-1)" and similar. That's not particularly helpful to me, so I added:
set mapred.job.name = more helpful name;
to the query. Now I can tell them apart better. However, now my queries that get split into multiple stages all show up as the same name. What I'd ideally like is something along the lines of set mapred.job.name = more helpful name (Stage-%d);
where the %d would get replaced by the current stage number.
Is this possible, and does anyone know how?
相关问题
- Spark on Yarn Container Failure
- enableHiveSupport throws error in java spark code
- spark select and add columns with alias
- Unable to generate jar file for Hadoop
-
hive: cast array
> into map
相关文章
- 在hive sql里怎么把"2020-10-26T08:41:19.000Z"这个字符串转换成年月日
- Java写文件至HDFS失败
- mapreduce count example
- SQL query Frequency Distribution matrix for produc
- Cloudera 5.6: Parquet does not support date. See H
- Could you give me any clue Why 'Cannot call me
- converting to timestamp with time zone failed on A
- Hive error: parseexception missing EOF
I'm not sure there is a way to implement exactly what you wish but I can offer something else.
Instead of using
set mapred.job.name
you can add a comment in the beginning of the query with amore helpful name
like this :-- this is a more helpful name
SELECT field, other_field ....
Then, in the jobtracker you'll see
-- this is a more helpful name ..... (Stage-%d)
"I've found this site: https://cwiki.apache.org/confluence/display/Hive/AdminManual+Configuration
on it there is a property called:
hive.query.string
so
set hive.query.string = even more helpful name
should work.It works perfectly for me.
I know this is a very late reply but anyways if this helps let me know.
This happens because HIVE does not allow certain parameters to be set at run time. Still if you want to set it follow the following steps:
KEY: hive.security.authorization.sqlstd.confwhitelist.append
VALUE: mapred.job.name
You can any key-value pair in this config for which you get this runtime error