|
23 | 23 | import org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe; |
24 | 24 | import org.apache.hadoop.hive.serde2.io.DateWritable; |
25 | 25 | import org.apache.hadoop.hive.serde2.io.HiveDecimalWritable; |
| 26 | +import org.apache.hadoop.hive.serde2.io.TimestampWritable; |
26 | 27 | import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; |
27 | 28 | import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorFactory; |
| 29 | +import org.apache.hadoop.io.BooleanWritable; |
28 | 30 | import org.apache.hadoop.io.ByteWritable; |
29 | 31 | import org.apache.hadoop.io.BytesWritable; |
30 | 32 | import org.apache.hadoop.io.DoubleWritable; |
| 33 | +import org.apache.hadoop.io.FloatWritable; |
31 | 34 | import org.apache.hadoop.io.IntWritable; |
32 | 35 | import org.apache.hadoop.io.LongWritable; |
33 | 36 | import org.apache.hadoop.io.Text; |
@@ -68,10 +71,17 @@ public static Object getWritableValue(Object writable) { |
68 | 71 | ret = ((DateWritable) writable).get(); |
69 | 72 | } else if(writable instanceof DoubleWritable){ |
70 | 73 | ret = ((DoubleWritable) writable).get(); |
71 | | - } else { |
| 74 | + } else if(writable instanceof TimestampWritable){ |
| 75 | + ret = ((TimestampWritable) writable).getTimestamp(); |
| 76 | + }else if (writable instanceof DateWritable){ |
| 77 | + ret = ((DateWritable) writable).get(); |
| 78 | + }else if (writable instanceof FloatWritable){ |
| 79 | + ret = ((FloatWritable) writable).get(); |
| 80 | + }else if (writable instanceof BooleanWritable){ |
| 81 | + ret = ((BooleanWritable) writable).get(); |
| 82 | + }else { |
72 | 83 | ret = writable.toString(); |
73 | 84 | } |
74 | | - |
75 | 85 | return ret; |
76 | 86 | } |
77 | 87 |
|
|
0 commit comments