@@ -931,11 +931,17 @@ func flattenHiveJob(job *dataproc.HiveJob) []map[string]interface{} {
931
931
func expandHiveJob (config map [string ]interface {}) * dataproc.HiveJob {
932
932
job := & dataproc.HiveJob {}
933
933
if v , ok := config ["query_file_uri" ]; ok {
934
- job .QueryFileUri = v .(string )
934
+ queryFileUri := v .(string )
935
+ if len (queryFileUri ) != 0 {
936
+ job .QueryFileUri = v .(string )
937
+ }
935
938
}
936
939
if v , ok := config ["query_list" ]; ok {
937
- job .QueryList = & dataproc.QueryList {
938
- Queries : tpgresource .ConvertStringArr (v .([]interface {})),
940
+ queryList := v .([]interface {})
941
+ if len (queryList ) != 0 {
942
+ job .QueryList = & dataproc.QueryList {
943
+ Queries : tpgresource .ConvertStringArr (queryList ),
944
+ }
939
945
}
940
946
}
941
947
if v , ok := config ["continue_on_failure" ]; ok {
@@ -1039,11 +1045,17 @@ func flattenPigJob(job *dataproc.PigJob) []map[string]interface{} {
1039
1045
func expandPigJob (config map [string ]interface {}) * dataproc.PigJob {
1040
1046
job := & dataproc.PigJob {}
1041
1047
if v , ok := config ["query_file_uri" ]; ok {
1042
- job .QueryFileUri = v .(string )
1048
+ queryFileUri := v .(string )
1049
+ if len (queryFileUri ) != 0 {
1050
+ job .QueryFileUri = v .(string )
1051
+ }
1043
1052
}
1044
1053
if v , ok := config ["query_list" ]; ok {
1045
- job .QueryList = & dataproc.QueryList {
1046
- Queries : tpgresource .ConvertStringArr (v .([]interface {})),
1054
+ queryList := v .([]interface {})
1055
+ if len (queryList ) != 0 {
1056
+ job .QueryList = & dataproc.QueryList {
1057
+ Queries : tpgresource .ConvertStringArr (queryList ),
1058
+ }
1047
1059
}
1048
1060
}
1049
1061
if v , ok := config ["continue_on_failure" ]; ok {
@@ -1140,11 +1152,17 @@ func flattenSparkSqlJob(job *dataproc.SparkSqlJob) []map[string]interface{} {
1140
1152
func expandSparkSqlJob (config map [string ]interface {}) * dataproc.SparkSqlJob {
1141
1153
job := & dataproc.SparkSqlJob {}
1142
1154
if v , ok := config ["query_file_uri" ]; ok {
1143
- job .QueryFileUri = v .(string )
1155
+ queryFileUri := v .(string )
1156
+ if len (queryFileUri ) != 0 {
1157
+ job .QueryFileUri = v .(string )
1158
+ }
1144
1159
}
1145
1160
if v , ok := config ["query_list" ]; ok {
1146
- job .QueryList = & dataproc.QueryList {
1147
- Queries : tpgresource .ConvertStringArr (v .([]interface {})),
1161
+ queryList := v .([]interface {})
1162
+ if len (queryList ) != 0 {
1163
+ job .QueryList = & dataproc.QueryList {
1164
+ Queries : tpgresource .ConvertStringArr (queryList ),
1165
+ }
1148
1166
}
1149
1167
}
1150
1168
if v , ok := config ["script_variables" ]; ok {
@@ -1241,20 +1259,26 @@ func flattenPrestoJob(job *dataproc.PrestoJob) []map[string]interface{} {
1241
1259
1242
1260
func expandPrestoJob (config map [string ]interface {}) * dataproc.PrestoJob {
1243
1261
job := & dataproc.PrestoJob {}
1262
+ if v , ok := config ["query_file_uri" ]; ok {
1263
+ queryFileUri := v .(string )
1264
+ if len (queryFileUri ) != 0 {
1265
+ job .QueryFileUri = v .(string )
1266
+ }
1267
+ }
1268
+ if v , ok := config ["query_list" ]; ok {
1269
+ queryList := v .([]interface {})
1270
+ if len (queryList ) != 0 {
1271
+ job .QueryList = & dataproc.QueryList {
1272
+ Queries : tpgresource .ConvertStringArr (queryList ),
1273
+ }
1274
+ }
1275
+ }
1244
1276
if v , ok := config ["client_tags" ]; ok {
1245
1277
job .ClientTags = tpgresource .ConvertStringArr (v .([]interface {}))
1246
1278
}
1247
1279
if v , ok := config ["continue_on_failure" ]; ok {
1248
1280
job .ContinueOnFailure = v .(bool )
1249
1281
}
1250
- if v , ok := config ["query_file_uri" ]; ok {
1251
- job .QueryFileUri = v .(string )
1252
- }
1253
- if v , ok := config ["query_list" ]; ok {
1254
- job .QueryList = & dataproc.QueryList {
1255
- Queries : tpgresource .ConvertStringArr (v .([]interface {})),
1256
- }
1257
- }
1258
1282
if v , ok := config ["properties" ]; ok {
1259
1283
job .Properties = tpgresource .ConvertStringMap (v .(map [string ]interface {}))
1260
1284
}
0 commit comments