diff --git a/api/src/main/java/io/druid/timeline/DataSegmentUtils.java b/api/src/main/java/io/druid/timeline/DataSegmentUtils.java index 349bbceda03..546b89dbd0e 100644 --- a/api/src/main/java/io/druid/timeline/DataSegmentUtils.java +++ b/api/src/main/java/io/druid/timeline/DataSegmentUtils.java @@ -21,13 +21,12 @@ package io.druid.timeline; import com.google.common.base.Function; import io.druid.guice.annotations.PublicApi; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.IAE; import io.druid.java.util.common.StringUtils; import io.druid.java.util.common.logger.Logger; import org.joda.time.DateTime; import org.joda.time.Interval; -import org.joda.time.format.DateTimeFormatter; -import org.joda.time.format.ISODateTimeFormat; import java.util.Objects; @@ -83,11 +82,9 @@ public class DataSegmentUtils return null; } - DateTimeFormatter formatter = ISODateTimeFormat.dateTime(); - try { - DateTime start = formatter.parseDateTime(splits[0]); - DateTime end = formatter.parseDateTime(splits[1]); + DateTime start = DateTimes.ISO_DATE_TIME.parse(splits[0]); + DateTime end = DateTimes.ISO_DATE_TIME.parse(splits[1]); String version = splits[2]; String trail = splits.length > 3 ? join(splits, DataSegment.delimiter, 3, splits.length) : null; diff --git a/api/src/test/java/io/druid/data/input/impl/TimestampSpecTest.java b/api/src/test/java/io/druid/data/input/impl/TimestampSpecTest.java index 52667a098b5..9aca31c4516 100644 --- a/api/src/test/java/io/druid/data/input/impl/TimestampSpecTest.java +++ b/api/src/test/java/io/druid/data/input/impl/TimestampSpecTest.java @@ -61,9 +61,11 @@ public class TimestampSpecTest }; TimestampSpec spec = new TimestampSpec("TIMEstamp", DATE_FORMAT, null); + DateTimes.UtcFormatter formatter = DateTimes.wrapFormatter(ISODateTimeFormat.dateHourMinuteSecond()); + for (String date : dates) { DateTime dateTime = spec.extractTimestamp(ImmutableMap.of("TIMEstamp", date)); - DateTime expectedDateTime = ISODateTimeFormat.dateHourMinuteSecond().parseDateTime(date); + DateTime expectedDateTime = formatter.parse(date); Assert.assertEquals(expectedDateTime, dateTime); } } diff --git a/codestyle/joda-time-forbidden-apis.txt b/codestyle/joda-time-forbidden-apis.txt index 76b1fe59f42..ae1ad41486c 100644 --- a/codestyle/joda-time-forbidden-apis.txt +++ b/codestyle/joda-time-forbidden-apis.txt @@ -1,4 +1,4 @@ -@defaultMessage Uses default time zone +@defaultMessage Uses default time zone, use factory methods from DateTimes and Intervals classes instead org.joda.time.DateTime#() org.joda.time.DateTime#(long) org.joda.time.DateTime#(java.lang.Object) @@ -6,11 +6,17 @@ org.joda.time.DateTime#(int, int, int, int, int) org.joda.time.DateTime#(int, int, int, int, int, int) org.joda.time.DateTime#(int, int, int, int, int, int, int) org.joda.time.DateTime#now() +org.joda.time.DateTime#parse(java.lang.String) +org.joda.time.DateTime#parse(java.lang.String, org.joda.time.format.DateTimeFormatter) +org.joda.time.base.BaseDateTime#() +org.joda.time.base.BaseDateTime#(long) +org.joda.time.base.BaseDateTime#(int, int, int, int, int, int, int) org.joda.time.Instant#toDateTime() +org.joda.time.Instant#toDateTimeISO() org.joda.time.Instant#toMutableDateTime() org.joda.time.Instant#toMutableDateTimeISO() -org.joda.time.base.AbstractInstant#toDateTimeISO() org.joda.time.base.AbstractInstant#toDateTime() +org.joda.time.base.AbstractInstant#toDateTimeISO() org.joda.time.base.AbstractInstant#toMutableDateTime() org.joda.time.base.AbstractInstant#toMutableDateTimeISO() org.joda.time.LocalDateTime#() @@ -40,6 +46,31 @@ org.joda.time.Interval#(long, long) org.joda.time.Interval#(java.lang.Object) org.joda.time.Interval#parse(java.lang.String) org.joda.time.Interval#parseWithOffset(java.lang.String) +org.joda.time.MutableDateTime#() +org.joda.time.MutableDateTime#(long) +org.joda.time.MutableDateTime#(java.lang.Object) +org.joda.time.MutableDateTime#(int, int, int, int, int, int, int) +org.joda.time.MutableDateTime#now() +org.joda.time.MutableDateTime#parse(java.lang.String) +org.joda.time.MutableDateTime#parse(java.lang.String, org.joda.time.format.DateTimeFormatter) +org.joda.time.MutableInterval#() +org.joda.time.MutableInterval#(long, long) +org.joda.time.MutableInterval#(java.lang.Object) +org.joda.time.MutableInterval#parse(java.lang.String) + +@defaultMessage Uses default time zone, use DateTimes.UtcFormatter to parse. +org.joda.time.format.DateTimeFormatter#parseInto(org.joda.time.ReadWritableInstant, java.lang.String, int) +org.joda.time.format.DateTimeFormatter#parseDateTime(java.lang.String) +org.joda.time.format.DateTimeFormatter#parseMutableDateTime(java.lang.String) + +@defaultMessage Do not use Chronology instance with default time zone. +org.joda.time.chrono.BuddhistChronology#getInstance() +org.joda.time.chrono.CopticChronology#getInstance() +org.joda.time.chrono.EthiopicChronology#getInstance() +org.joda.time.chrono.GJChronology#getInstance() +org.joda.time.chrono.GregorianChronology#getInstance() +org.joda.time.chrono.IslamicChronology#getInstance() +org.joda.time.chrono.ISOChronology#getInstance() @defaultMessage Doesn't handle edge cases where the start of day isn't midnight. org.joda.time.LocalDate#toDateTimeAtMidnight() diff --git a/common/src/main/java/io/druid/math/expr/Function.java b/common/src/main/java/io/druid/math/expr/Function.java index e1510a9647f..a788e21c569 100644 --- a/common/src/main/java/io/druid/math/expr/Function.java +++ b/common/src/main/java/io/druid/math/expr/Function.java @@ -20,12 +20,11 @@ package io.druid.math.expr; import com.google.common.base.Strings; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.IAE; import io.druid.java.util.common.StringUtils; import org.joda.time.DateTime; import org.joda.time.format.DateTimeFormat; -import org.joda.time.format.DateTimeFormatter; -import org.joda.time.format.ISODateTimeFormat; import java.util.List; @@ -824,17 +823,17 @@ interface Function throw new IAE("first argument should be string type but got %s type", value.type()); } - DateTimeFormatter formatter = ISODateTimeFormat.dateOptionalTimeParser(); + DateTimes.UtcFormatter formatter = DateTimes.ISO_DATE_OPTIONAL_TIME; if (args.size() > 1) { ExprEval format = args.get(1).eval(bindings); if (format.type() != ExprType.STRING) { throw new IAE("second argument should be string type but got %s type", format.type()); } - formatter = DateTimeFormat.forPattern(format.asString()); + formatter = DateTimes.wrapFormatter(DateTimeFormat.forPattern(format.asString())); } DateTime date; try { - date = DateTime.parse(value.asString(), formatter); + date = formatter.parse(value.asString()); } catch (IllegalArgumentException e) { throw new IAE(e, "invalid value %s", value.asString()); diff --git a/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampAggregationSelectTest.java b/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampAggregationSelectTest.java index 91483537761..1e45410ae87 100644 --- a/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampAggregationSelectTest.java +++ b/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampAggregationSelectTest.java @@ -25,6 +25,7 @@ import com.google.common.collect.ImmutableList; import com.google.common.collect.Iterables; import com.google.common.collect.Lists; import com.google.common.io.Resources; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.granularity.Granularities; import io.druid.java.util.common.guava.Sequence; import io.druid.java.util.common.guava.Sequences; @@ -32,7 +33,6 @@ import io.druid.query.Result; import io.druid.query.select.SelectResultValue; import io.druid.segment.ColumnSelectorFactory; import org.easymock.EasyMock; -import org.joda.time.DateTime; import org.junit.Assert; import org.junit.Before; import org.junit.Rule; @@ -65,8 +65,8 @@ public class TimestampAggregationSelectTest { return Iterables.transform( ImmutableList.of( - ImmutableList.of("timeMin", "tmin", TimestampMinAggregatorFactory.class, DateTime.parse("2011-01-12T01:00:00.000Z").getMillis()), - ImmutableList.of("timeMax", "tmax", TimestampMaxAggregatorFactory.class, DateTime.parse("2011-01-31T01:00:00.000Z").getMillis()) + ImmutableList.of("timeMin", "tmin", TimestampMinAggregatorFactory.class, DateTimes.of("2011-01-12T01:00:00.000Z").getMillis()), + ImmutableList.of("timeMax", "tmax", TimestampMaxAggregatorFactory.class, DateTimes.of("2011-01-31T01:00:00.000Z").getMillis()) ), new Function, Object[]>() { diff --git a/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampGroupByAggregationTest.java b/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampGroupByAggregationTest.java index c192c103202..f35133715fd 100644 --- a/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampGroupByAggregationTest.java +++ b/extensions-contrib/time-min-max/src/test/java/io/druid/query/aggregation/TimestampGroupByAggregationTest.java @@ -23,6 +23,7 @@ import com.google.common.collect.ImmutableList; import com.google.common.collect.Lists; import io.druid.data.input.MapBasedRow; import io.druid.data.input.Row; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.granularity.Granularities; import io.druid.java.util.common.guava.Sequence; import io.druid.java.util.common.guava.Sequences; @@ -63,8 +64,8 @@ public class TimestampGroupByAggregationTest final List constructors = Lists.newArrayList(); final List> partialConstructors = ImmutableList.>of( - ImmutableList.of("timeMin", "tmin", "time_min", TimestampMinAggregatorFactory.class, DateTime.parse("2011-01-12T01:00:00.000Z")), - ImmutableList.of("timeMax", "tmax", "time_max", TimestampMaxAggregatorFactory.class, DateTime.parse("2011-01-31T01:00:00.000Z")) + ImmutableList.of("timeMin", "tmin", "time_min", TimestampMinAggregatorFactory.class, DateTimes.of("2011-01-12T01:00:00.000Z")), + ImmutableList.of("timeMax", "tmax", "time_max", TimestampMaxAggregatorFactory.class, DateTimes.of("2011-01-31T01:00:00.000Z")) ); for (final List partialConstructor : partialConstructors) { diff --git a/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationTest.java b/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationTest.java index 49352ac18f1..c5d302651ce 100644 --- a/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationTest.java +++ b/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationTest.java @@ -31,6 +31,7 @@ import com.yahoo.sketches.theta.Union; import com.yahoo.sketches.theta.UpdateSketch; import io.druid.data.input.MapBasedRow; import io.druid.data.input.Row; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.granularity.Granularities; import io.druid.java.util.common.guava.Sequence; import io.druid.java.util.common.guava.Sequences; @@ -42,7 +43,6 @@ import io.druid.query.groupby.GroupByQueryConfig; import io.druid.query.groupby.GroupByQueryRunnerTest; import io.druid.query.groupby.epinephelinae.GrouperTestUtil; import io.druid.query.groupby.epinephelinae.TestColumnSelectorFactory; -import org.joda.time.DateTime; import org.junit.Assert; import org.junit.Rule; import org.junit.Test; @@ -106,7 +106,7 @@ public class SketchAggregationTest Assert.assertEquals(1, results.size()); Assert.assertEquals( new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("sids_sketch_count", 50.0) @@ -148,7 +148,7 @@ public class SketchAggregationTest Assert.assertEquals( ImmutableList.of( new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_3") @@ -161,7 +161,7 @@ public class SketchAggregationTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_1") @@ -174,7 +174,7 @@ public class SketchAggregationTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_2") @@ -187,7 +187,7 @@ public class SketchAggregationTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_4") @@ -200,7 +200,7 @@ public class SketchAggregationTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_5") @@ -346,7 +346,7 @@ public class SketchAggregationTest Assert.assertEquals( ImmutableList.of( new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_1") diff --git a/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationWithSimpleDataTest.java b/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationWithSimpleDataTest.java index 2583229d6ef..00fd0b569d0 100644 --- a/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationWithSimpleDataTest.java +++ b/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/SketchAggregationWithSimpleDataTest.java @@ -38,7 +38,6 @@ import io.druid.query.select.SelectResultValue; import io.druid.query.timeseries.TimeseriesResultValue; import io.druid.query.topn.DimensionAndMetricValueExtractor; import io.druid.query.topn.TopNResultValue; -import org.joda.time.DateTime; import org.junit.Assert; import org.junit.Before; import org.junit.Rule; @@ -136,7 +135,7 @@ public class SketchAggregationWithSimpleDataTest Assert.assertEquals( ImmutableList.of( new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_3") @@ -149,7 +148,7 @@ public class SketchAggregationWithSimpleDataTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_1") @@ -162,7 +161,7 @@ public class SketchAggregationWithSimpleDataTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_2") @@ -175,7 +174,7 @@ public class SketchAggregationWithSimpleDataTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_4") @@ -188,7 +187,7 @@ public class SketchAggregationWithSimpleDataTest .build() ), new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("product", "product_5") diff --git a/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/oldapi/OldApiSketchAggregationTest.java b/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/oldapi/OldApiSketchAggregationTest.java index a001ed77f7a..fdabf39c6fc 100644 --- a/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/oldapi/OldApiSketchAggregationTest.java +++ b/extensions-core/datasketches/src/test/java/io/druid/query/aggregation/datasketches/theta/oldapi/OldApiSketchAggregationTest.java @@ -25,6 +25,7 @@ import com.google.common.io.Files; import com.yahoo.sketches.theta.Sketches; import com.yahoo.sketches.theta.UpdateSketch; import io.druid.data.input.MapBasedRow; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.granularity.Granularities; import io.druid.java.util.common.guava.Sequence; import io.druid.java.util.common.guava.Sequences; @@ -37,7 +38,6 @@ import io.druid.query.groupby.GroupByQueryConfig; import io.druid.query.groupby.GroupByQueryRunnerTest; import io.druid.query.groupby.epinephelinae.GrouperTestUtil; import io.druid.query.groupby.epinephelinae.TestColumnSelectorFactory; -import org.joda.time.DateTime; import org.junit.Assert; import org.junit.Rule; import org.junit.Test; @@ -100,7 +100,7 @@ public class OldApiSketchAggregationTest Assert.assertEquals(1, results.size()); Assert.assertEquals( new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("sketch_count", 50.0) @@ -132,7 +132,7 @@ public class OldApiSketchAggregationTest Assert.assertEquals(1, results.size()); Assert.assertEquals( new MapBasedRow( - DateTime.parse("2014-10-19T00:00:00.000Z"), + DateTimes.of("2014-10-19T00:00:00.000Z"), ImmutableMap .builder() .put("sids_sketch_count", 50.0) diff --git a/extensions-core/kafka-extraction-namespace/src/test/java/io/druid/query/lookup/TestKafkaExtractionCluster.java b/extensions-core/kafka-extraction-namespace/src/test/java/io/druid/query/lookup/TestKafkaExtractionCluster.java index 6c9ad105417..bdc421b43f0 100644 --- a/extensions-core/kafka-extraction-namespace/src/test/java/io/druid/query/lookup/TestKafkaExtractionCluster.java +++ b/extensions-core/kafka-extraction-namespace/src/test/java/io/druid/query/lookup/TestKafkaExtractionCluster.java @@ -28,6 +28,7 @@ import com.google.inject.Module; import com.google.inject.name.Names; import io.druid.guice.GuiceInjectors; import io.druid.initialization.Initialization; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.ISE; import io.druid.java.util.common.StringUtils; import io.druid.java.util.common.io.Closer; @@ -45,7 +46,6 @@ import org.I0Itec.zkclient.ZkClient; import org.I0Itec.zkclient.exception.ZkException; import org.apache.curator.test.TestingServer; import org.apache.zookeeper.CreateMode; -import org.joda.time.DateTime; import org.junit.After; import org.junit.Assert; import org.junit.Before; @@ -134,7 +134,7 @@ public class TestKafkaExtractionCluster kafkaConfig = new KafkaConfig(serverProperties); - final long time = DateTime.parse("2015-01-01").getMillis(); + final long time = DateTimes.of("2015-01-01").getMillis(); kafkaServer = new KafkaServer( kafkaConfig, new Time() diff --git a/indexing-hadoop/src/test/java/io/druid/indexer/BatchDeltaIngestionTest.java b/indexing-hadoop/src/test/java/io/druid/indexer/BatchDeltaIngestionTest.java index 86cfc9ca59e..1834415036e 100644 --- a/indexing-hadoop/src/test/java/io/druid/indexer/BatchDeltaIngestionTest.java +++ b/indexing-hadoop/src/test/java/io/druid/indexer/BatchDeltaIngestionTest.java @@ -35,6 +35,7 @@ import io.druid.data.input.impl.StringInputRowParser; import io.druid.data.input.impl.TimestampSpec; import io.druid.indexer.hadoop.WindowedDataSegment; import io.druid.jackson.DefaultObjectMapper; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.Intervals; import io.druid.java.util.common.StringUtils; import io.druid.java.util.common.granularity.Granularities; @@ -53,7 +54,6 @@ import io.druid.segment.realtime.firehose.WindowedStorageAdapter; import io.druid.timeline.DataSegment; import io.druid.timeline.partition.HashBasedNumberedShardSpec; import org.apache.commons.io.FileUtils; -import org.joda.time.DateTime; import org.joda.time.Interval; import org.junit.Assert; import org.junit.Rule; @@ -127,19 +127,19 @@ public class BatchDeltaIngestionTest List> expectedRows = ImmutableList.of( ImmutableMap.of( - "time", DateTime.parse("2014-10-22T00:00:00.000Z"), + "time", DateTimes.of("2014-10-22T00:00:00.000Z"), "host", ImmutableList.of("a.example.com"), "visited_sum", 100L, "unique_hosts", 1.0d ), ImmutableMap.of( - "time", DateTime.parse("2014-10-22T01:00:00.000Z"), + "time", DateTimes.of("2014-10-22T01:00:00.000Z"), "host", ImmutableList.of("b.example.com"), "visited_sum", 150L, "unique_hosts", 1.0d ), ImmutableMap.of( - "time", DateTime.parse("2014-10-22T02:00:00.000Z"), + "time", DateTimes.of("2014-10-22T02:00:00.000Z"), "host", ImmutableList.of("c.example.com"), "visited_sum", 200L, "unique_hosts", 1.0d @@ -173,13 +173,13 @@ public class BatchDeltaIngestionTest List> expectedRows = ImmutableList.of( ImmutableMap.of( - "time", DateTime.parse("2014-10-22T00:00:00.000Z"), + "time", DateTimes.of("2014-10-22T00:00:00.000Z"), "host", ImmutableList.of("a.example.com"), "visited_sum", 100L, "unique_hosts", 1.0d ), ImmutableMap.of( - "time", DateTime.parse("2014-10-22T01:00:00.000Z"), + "time", DateTimes.of("2014-10-22T01:00:00.000Z"), "host", ImmutableList.of("b.example.com"), "visited_sum", 150L, "unique_hosts", 1.0d @@ -250,19 +250,19 @@ public class BatchDeltaIngestionTest List> expectedRows = ImmutableList.of( ImmutableMap.of( - "time", DateTime.parse("2014-10-22T00:00:00.000Z"), + "time", DateTimes.of("2014-10-22T00:00:00.000Z"), "host", ImmutableList.of("a.example.com"), "visited_sum", 190L, "unique_hosts", 1.0d ), ImmutableMap.of( - "time", DateTime.parse("2014-10-22T01:00:00.000Z"), + "time", DateTimes.of("2014-10-22T01:00:00.000Z"), "host", ImmutableList.of("b.example.com"), "visited_sum", 175L, "unique_hosts", 1.0d ), ImmutableMap.of( - "time", DateTime.parse("2014-10-22T02:00:00.000Z"), + "time", DateTimes.of("2014-10-22T02:00:00.000Z"), "host", ImmutableList.of("c.example.com"), "visited_sum", 270L, "unique_hosts", 1.0d diff --git a/indexing-hadoop/src/test/java/io/druid/indexer/hadoop/DatasourceRecordReaderTest.java b/indexing-hadoop/src/test/java/io/druid/indexer/hadoop/DatasourceRecordReaderTest.java index 50aa49fa827..fd6c03bb67c 100644 --- a/indexing-hadoop/src/test/java/io/druid/indexer/hadoop/DatasourceRecordReaderTest.java +++ b/indexing-hadoop/src/test/java/io/druid/indexer/hadoop/DatasourceRecordReaderTest.java @@ -25,13 +25,13 @@ import com.google.common.collect.Lists; import io.druid.data.input.InputRow; import io.druid.indexer.HadoopDruidIndexerConfig; import io.druid.jackson.DefaultObjectMapper; +import io.druid.java.util.common.DateTimes; import io.druid.query.aggregation.hyperloglog.HyperUniquesAggregatorFactory; import io.druid.timeline.DataSegment; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.mapreduce.InputSplit; import org.apache.hadoop.mapreduce.TaskAttemptContext; import org.easymock.EasyMock; -import org.joda.time.DateTime; import org.junit.Assert; import org.junit.Test; @@ -98,19 +98,19 @@ public class DatasourceRecordReaderTest { List> expectedRows = ImmutableList.of( ImmutableMap.of( - "time", DateTime.parse("2014-10-22T00:00:00.000Z"), + "time", DateTimes.of("2014-10-22T00:00:00.000Z"), "host", ImmutableList.of("a.example.com"), "visited_sum", 100L, "unique_hosts", 1.0d ), ImmutableMap.of( - "time", DateTime.parse("2014-10-22T01:00:00.000Z"), + "time", DateTimes.of("2014-10-22T01:00:00.000Z"), "host", ImmutableList.of("b.example.com"), "visited_sum", 150L, "unique_hosts", 1.0d ), ImmutableMap.of( - "time", DateTime.parse("2014-10-22T02:00:00.000Z"), + "time", DateTimes.of("2014-10-22T02:00:00.000Z"), "host", ImmutableList.of("c.example.com"), "visited_sum", 200L, "unique_hosts", 1.0d diff --git a/java-util/src/main/java/io/druid/java/util/common/DateTimes.java b/java-util/src/main/java/io/druid/java/util/common/DateTimes.java index 149e34ee397..e5a08595c25 100644 --- a/java-util/src/main/java/io/druid/java/util/common/DateTimes.java +++ b/java-util/src/main/java/io/druid/java/util/common/DateTimes.java @@ -19,8 +19,11 @@ package io.druid.java.util.common; +import org.joda.time.Chronology; import org.joda.time.DateTime; import org.joda.time.chrono.ISOChronology; +import org.joda.time.format.DateTimeFormatter; +import org.joda.time.format.ISODateTimeFormat; public final class DateTimes { @@ -28,6 +31,42 @@ public final class DateTimes public static final DateTime MAX = utc(JodaUtils.MAX_INSTANT); public static final DateTime MIN = utc(JodaUtils.MIN_INSTANT); + public static final UtcFormatter ISO_DATE_TIME = wrapFormatter(ISODateTimeFormat.dateTime()); + public static final UtcFormatter ISO_DATE_OPTIONAL_TIME = wrapFormatter(ISODateTimeFormat.dateOptionalTimeParser()); + public static final UtcFormatter ISO_DATE_OR_TIME = wrapFormatter(ISODateTimeFormat.dateTimeParser()); + public static final UtcFormatter ISO_DATE_OR_TIME_WITH_OFFSET = wrapFormatter( + ISODateTimeFormat.dateTimeParser().withOffsetParsed() + ); + + /** + * Simple wrapper class to enforce UTC Chronology in formatter. Specifically, it will use + * {@link DateTimeFormatter#withChronology(Chronology)} to set the chronology to + * {@link ISOChronology#getInstanceUTC()} on the wrapped {@link DateTimeFormatter}. + */ + public static class UtcFormatter + { + private final DateTimeFormatter innerFormatter; + + private UtcFormatter(final DateTimeFormatter innerFormatter) + { + this.innerFormatter = innerFormatter.withChronology(ISOChronology.getInstanceUTC()); + } + + public DateTime parse(final String instant) + { + return innerFormatter.parseDateTime(instant); + } + } + + /** + * Creates a {@link UtcFormatter} that wraps around a {@link DateTimeFormatter}. + * @param formatter inner {@link DateTimeFormatter} used to parse {@link String} + */ + public static UtcFormatter wrapFormatter(final DateTimeFormatter formatter) + { + return new UtcFormatter(formatter); + } + public static DateTime utc(long instant) { return new DateTime(instant, ISOChronology.getInstanceUTC()); diff --git a/java-util/src/main/java/io/druid/java/util/common/parsers/TimestampParser.java b/java-util/src/main/java/io/druid/java/util/common/parsers/TimestampParser.java index 58b5af67f3f..5624f393348 100644 --- a/java-util/src/main/java/io/druid/java/util/common/parsers/TimestampParser.java +++ b/java-util/src/main/java/io/druid/java/util/common/parsers/TimestampParser.java @@ -42,7 +42,7 @@ public class TimestampParser { if (format.equalsIgnoreCase("auto")) { // Could be iso or millis - final DateTimeFormatter parser = createAutoParser(); + final DateTimes.UtcFormatter parser = DateTimes.wrapFormatter(createAutoParser()); return (String input) -> { Preconditions.checkArgument(!Strings.isNullOrEmpty(input), "null timestamp"); @@ -59,7 +59,7 @@ public class TimestampParser } } - return new DateTime(parser.parseDateTime(input), timeZone); + return parser.parse(input).withZone(timeZone); } } @@ -87,10 +87,10 @@ public class TimestampParser }; } else { try { - final DateTimeFormatter formatter = DateTimeFormat.forPattern(format); + final DateTimes.UtcFormatter formatter = DateTimes.wrapFormatter(DateTimeFormat.forPattern(format)); return input -> { Preconditions.checkArgument(!Strings.isNullOrEmpty(input), "null timestamp"); - return formatter.parseDateTime(ParserUtils.stripQuotes(input)); + return formatter.parse(ParserUtils.stripQuotes(input)); }; } catch (Exception e) { diff --git a/pom.xml b/pom.xml index daa0fe06f7b..2fb94656c4b 100644 --- a/pom.xml +++ b/pom.xml @@ -940,6 +940,9 @@ ${session.executionRootDirectory}/codestyle/joda-time-forbidden-apis.txt ${session.executionRootDirectory}/codestyle/druid-forbidden-apis.txt + + io/druid/java/util/common/DateTimes$UtcFormatter.class + diff --git a/processing/src/main/java/io/druid/jackson/JodaStuff.java b/processing/src/main/java/io/druid/jackson/JodaStuff.java index 88af3c4dbcd..4ea8ae9061c 100644 --- a/processing/src/main/java/io/druid/jackson/JodaStuff.java +++ b/processing/src/main/java/io/druid/jackson/JodaStuff.java @@ -36,7 +36,6 @@ import org.joda.time.DateTime; import org.joda.time.Duration; import org.joda.time.Interval; import org.joda.time.Period; -import org.joda.time.format.ISODateTimeFormat; import java.io.IOException; @@ -108,9 +107,7 @@ class JodaStuff return null; } // make sure to preserve time zone information when parsing timestamps - return ISODateTimeFormat.dateTimeParser() - .withOffsetParsed() - .parseDateTime(str); + return DateTimes.ISO_DATE_OR_TIME_WITH_OFFSET.parse(str); } throw ctxt.mappingException(getValueClass()); } diff --git a/processing/src/main/java/io/druid/query/expression/TimestampParseExprMacro.java b/processing/src/main/java/io/druid/query/expression/TimestampParseExprMacro.java index b3a5ec829eb..ebc4ddecb83 100644 --- a/processing/src/main/java/io/druid/query/expression/TimestampParseExprMacro.java +++ b/processing/src/main/java/io/druid/query/expression/TimestampParseExprMacro.java @@ -19,14 +19,13 @@ package io.druid.query.expression; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.IAE; import io.druid.math.expr.Expr; import io.druid.math.expr.ExprEval; import io.druid.math.expr.ExprMacroTable; import org.joda.time.DateTimeZone; import org.joda.time.format.DateTimeFormat; -import org.joda.time.format.DateTimeFormatter; -import org.joda.time.format.ISODateTimeFormat; import javax.annotation.Nonnull; import java.util.List; @@ -56,9 +55,10 @@ public class TimestampParseExprMacro implements ExprMacroTable.ExprMacro timeZone = DateTimeZone.UTC; } - final DateTimeFormatter formatter = formatString == null - ? ISODateTimeFormat.dateTimeParser() - : DateTimeFormat.forPattern(formatString).withZone(timeZone); + final DateTimes.UtcFormatter formatter = + formatString == null + ? DateTimes.ISO_DATE_OR_TIME + : DateTimes.wrapFormatter(DateTimeFormat.forPattern(formatString).withZone(timeZone)); class TimestampParseExpr implements Expr { @@ -67,7 +67,7 @@ public class TimestampParseExprMacro implements ExprMacroTable.ExprMacro public ExprEval eval(final ObjectBinding bindings) { try { - return ExprEval.of(formatter.parseDateTime(arg.eval(bindings).asString()).getMillis()); + return ExprEval.of(formatter.parse(arg.eval(bindings).asString()).getMillis()); } catch (IllegalArgumentException e) { // Catch exceptions potentially thrown by formatter.parseDateTime. Our docs say that unparseable timestamps diff --git a/processing/src/test/java/io/druid/segment/IndexMergerV9CompatibilityTest.java b/processing/src/test/java/io/druid/segment/IndexMergerV9CompatibilityTest.java index f116a2c022a..060bca51e44 100644 --- a/processing/src/test/java/io/druid/segment/IndexMergerV9CompatibilityTest.java +++ b/processing/src/test/java/io/druid/segment/IndexMergerV9CompatibilityTest.java @@ -23,9 +23,10 @@ import com.google.common.collect.ImmutableList; import com.google.common.collect.ImmutableMap; import com.google.common.io.ByteSource; import com.google.common.io.Files; -import io.druid.java.util.common.JodaUtils; import io.druid.data.input.InputRow; import io.druid.data.input.MapBasedInputRow; +import io.druid.java.util.common.DateTimes; +import io.druid.java.util.common.JodaUtils; import io.druid.query.aggregation.AggregatorFactory; import io.druid.query.aggregation.CountAggregatorFactory; import io.druid.segment.data.CompressedObjectStrategy; @@ -34,7 +35,6 @@ import io.druid.segment.data.ConciseBitmapSerdeFactory; import io.druid.segment.incremental.IncrementalIndex; import io.druid.segment.incremental.IncrementalIndexSchema; import org.apache.commons.io.FileUtils; -import org.joda.time.DateTime; import org.junit.After; import org.junit.Assert; import org.junit.Before; @@ -56,7 +56,7 @@ public class IndexMergerV9CompatibilityTest { @Rule public final CloserRule closer = new CloserRule(false); - private static final long TIMESTAMP = DateTime.parse("2014-01-01").getMillis(); + private static final long TIMESTAMP = DateTimes.of("2014-01-01").getMillis(); private static final AggregatorFactory[] DEFAULT_AGG_FACTORIES = new AggregatorFactory[]{ new CountAggregatorFactory( "count" diff --git a/server/src/main/java/io/druid/metadata/SQLMetadataSegmentManager.java b/server/src/main/java/io/druid/metadata/SQLMetadataSegmentManager.java index 7308e2a46cd..2f3b083571a 100644 --- a/server/src/main/java/io/druid/metadata/SQLMetadataSegmentManager.java +++ b/server/src/main/java/io/druid/metadata/SQLMetadataSegmentManager.java @@ -47,7 +47,6 @@ import io.druid.timeline.DataSegment; import io.druid.timeline.TimelineObjectHolder; import io.druid.timeline.VersionedIntervalTimeline; import io.druid.timeline.partition.PartitionChunk; -import org.joda.time.DateTime; import org.joda.time.Duration; import org.joda.time.Interval; import org.skife.jdbi.v2.BaseResultSetMapper; @@ -574,8 +573,8 @@ public class SQLMetadataSegmentManager implements MetadataSegmentManager protected Interval mapInternal(int index, Map row) { return new Interval( - DateTime.parse((String) row.get("start")), - DateTime.parse((String) row.get("end")) + DateTimes.of((String) row.get("start")), + DateTimes.of((String) row.get("end")) ); } } diff --git a/server/src/test/java/io/druid/server/log/LoggingRequestLoggerTest.java b/server/src/test/java/io/druid/server/log/LoggingRequestLoggerTest.java index 83e1d38e78a..3a7020808ad 100644 --- a/server/src/test/java/io/druid/server/log/LoggingRequestLoggerTest.java +++ b/server/src/test/java/io/druid/server/log/LoggingRequestLoggerTest.java @@ -24,6 +24,7 @@ import com.fasterxml.jackson.databind.ObjectMapper; import com.google.common.base.Charsets; import com.google.common.collect.ImmutableMap; import io.druid.jackson.DefaultObjectMapper; +import io.druid.java.util.common.DateTimes; import io.druid.java.util.common.Intervals; import io.druid.java.util.common.jackson.JacksonUtils; import io.druid.query.BaseQuery; @@ -63,7 +64,7 @@ public class LoggingRequestLoggerTest private static final ByteArrayOutputStream baos = new ByteArrayOutputStream(); private static Appender appender; - final DateTime timestamp = DateTime.parse("2016-01-01T00:00:00Z"); + final DateTime timestamp = DateTimes.of("2016-01-01T00:00:00Z"); final String remoteAddr = "some.host.tld"; final Map queryContext = ImmutableMap.of("foo", "bar"); final Query query = new FakeQuery(