@@ -65,6 +65,7 @@ private[spark] object HiveUtils extends Logging {
65
65
.doc(" Version of the Hive metastore. Available options are " +
66
66
" <code>0.12.0</code> through <code>2.3.6</code> and " +
67
67
" <code>3.0.0</code> through <code>3.1.2</code>." )
68
+ .version(" 1.4.0" )
68
69
.stringConf
69
70
.createWithDefault(builtinHiveVersion)
70
71
@@ -73,6 +74,7 @@ private[spark] object HiveUtils extends Logging {
73
74
// already rely on this config.
74
75
val FAKE_HIVE_VERSION = buildConf(" spark.sql.hive.version" )
75
76
.doc(s " deprecated, please use ${HIVE_METASTORE_VERSION .key} to get the Hive version in Spark. " )
77
+ .version(" 1.1.1" )
76
78
.stringConf
77
79
.createWithDefault(builtinHiveVersion)
78
80
@@ -89,12 +91,14 @@ private[spark] object HiveUtils extends Logging {
89
91
| Use Hive jars of specified version downloaded from Maven repositories.
90
92
| 3. A classpath in the standard format for both Hive and Hadoop.
91
93
""" .stripMargin)
94
+ .version(" 1.4.0" )
92
95
.stringConf
93
96
.createWithDefault(" builtin" )
94
97
95
98
val CONVERT_METASTORE_PARQUET = buildConf(" spark.sql.hive.convertMetastoreParquet" )
96
99
.doc(" When set to true, the built-in Parquet reader and writer are used to process " +
97
100
" parquet tables created by using the HiveQL syntax, instead of Hive serde." )
101
+ .version(" 1.1.1" )
98
102
.booleanConf
99
103
.createWithDefault(true )
100
104
@@ -103,12 +107,14 @@ private[spark] object HiveUtils extends Logging {
103
107
.doc(" When true, also tries to merge possibly different but compatible Parquet schemas in " +
104
108
" different Parquet data files. This configuration is only effective " +
105
109
" when \" spark.sql.hive.convertMetastoreParquet\" is true." )
110
+ .version(" 1.3.1" )
106
111
.booleanConf
107
112
.createWithDefault(false )
108
113
109
114
val CONVERT_METASTORE_ORC = buildConf(" spark.sql.hive.convertMetastoreOrc" )
110
115
.doc(" When set to true, the built-in ORC reader and writer are used to process " +
111
116
" ORC tables created by using the HiveQL syntax, instead of Hive serde." )
117
+ .version(" 2.0.0" )
112
118
.booleanConf
113
119
.createWithDefault(true )
114
120
@@ -118,6 +124,7 @@ private[spark] object HiveUtils extends Logging {
118
124
" `spark.sql.hive.convertMetastoreOrc` is true, the built-in ORC/Parquet writer is used" +
119
125
" to process inserting into partitioned ORC/Parquet tables created by using the HiveSQL " +
120
126
" syntax." )
127
+ .version(" 3.0.0" )
121
128
.booleanConf
122
129
.createWithDefault(true )
123
130
@@ -126,6 +133,7 @@ private[spark] object HiveUtils extends Logging {
126
133
" instead of Hive serde in CTAS. This flag is effective only if " +
127
134
" `spark.sql.hive.convertMetastoreParquet` or `spark.sql.hive.convertMetastoreOrc` is " +
128
135
" enabled respectively for Parquet and ORC formats" )
136
+ .version(" 3.0.0" )
129
137
.booleanConf
130
138
.createWithDefault(true )
131
139
@@ -135,6 +143,7 @@ private[spark] object HiveUtils extends Logging {
135
143
" that should be shared is JDBC drivers that are needed to talk to the metastore. Other " +
136
144
" classes that need to be shared are those that interact with classes that are already " +
137
145
" shared. For example, custom appenders that are used by log4j." )
146
+ .version(" 1.4.0" )
138
147
.stringConf
139
148
.toSequence
140
149
.createWithDefault(jdbcPrefixes)
@@ -146,12 +155,14 @@ private[spark] object HiveUtils extends Logging {
146
155
.doc(" A comma separated list of class prefixes that should explicitly be reloaded for each " +
147
156
" version of Hive that Spark SQL is communicating with. For example, Hive UDFs that are " +
148
157
" declared in a prefix that typically would be shared (i.e. <code>org.apache.spark.*</code>)." )
158
+ .version(" 1.4.0" )
149
159
.stringConf
150
160
.toSequence
151
161
.createWithDefault(Nil )
152
162
153
163
val HIVE_THRIFT_SERVER_ASYNC = buildConf(" spark.sql.hive.thriftServer.async" )
154
164
.doc(" When set to true, Hive Thrift server executes SQL queries in an asynchronous way." )
165
+ .version(" 1.5.0" )
155
166
.booleanConf
156
167
.createWithDefault(true )
157
168
0 commit comments