未验证 提交 513ae013 编写于 作者: wu-sheng's avatar wu-sheng 提交者: GitHub

Merge pull request #606 from OpenSkywalking/feature/582

Add TIME column for table segment data
...@@ -53,6 +53,7 @@ public class SegmentEsPersistenceDAO extends EsDAO implements ISegmentPersistenc ...@@ -53,6 +53,7 @@ public class SegmentEsPersistenceDAO extends EsDAO implements ISegmentPersistenc
@Override public IndexRequestBuilder prepareBatchInsert(Segment data) { @Override public IndexRequestBuilder prepareBatchInsert(Segment data) {
Map<String, Object> source = new HashMap<>(); Map<String, Object> source = new HashMap<>();
source.put(SegmentTable.COLUMN_DATA_BINARY, new String(Base64.getEncoder().encode(data.getDataBinary()))); source.put(SegmentTable.COLUMN_DATA_BINARY, new String(Base64.getEncoder().encode(data.getDataBinary())));
source.put(SegmentTable.COLUMN_TIME_BUCKET, data.getTimeBucket());
logger.debug("segment source: {}", source.toString()); logger.debug("segment source: {}", source.toString());
return getClient().prepareIndex(SegmentTable.TABLE, data.getId()).setSource(source); return getClient().prepareIndex(SegmentTable.TABLE, data.getId()).setSource(source);
} }
......
...@@ -37,5 +37,6 @@ public class SegmentEsTableDefine extends ElasticSearchTableDefine { ...@@ -37,5 +37,6 @@ public class SegmentEsTableDefine extends ElasticSearchTableDefine {
@Override public void initialize() { @Override public void initialize() {
addColumn(new ElasticSearchColumnDefine(SegmentTable.COLUMN_DATA_BINARY, ElasticSearchColumnDefine.Type.Binary.name())); addColumn(new ElasticSearchColumnDefine(SegmentTable.COLUMN_DATA_BINARY, ElasticSearchColumnDefine.Type.Binary.name()));
addColumn(new ElasticSearchColumnDefine(SegmentTable.COLUMN_TIME_BUCKET, ElasticSearchColumnDefine.Type.Long.name()));
} }
} }
...@@ -50,6 +50,7 @@ public class SegmentH2PersistenceDAO extends H2DAO implements ISegmentPersistenc ...@@ -50,6 +50,7 @@ public class SegmentH2PersistenceDAO extends H2DAO implements ISegmentPersistenc
H2SqlEntity entity = new H2SqlEntity(); H2SqlEntity entity = new H2SqlEntity();
source.put(SegmentTable.COLUMN_ID, data.getId()); source.put(SegmentTable.COLUMN_ID, data.getId());
source.put(SegmentTable.COLUMN_DATA_BINARY, data.getDataBinary()); source.put(SegmentTable.COLUMN_DATA_BINARY, data.getDataBinary());
source.put(SegmentTable.COLUMN_TIME_BUCKET, data.getTimeBucket());
logger.debug("segment source: {}", source.toString()); logger.debug("segment source: {}", source.toString());
String sql = SqlBuilder.buildBatchInsertSql(SegmentTable.TABLE, source.keySet()); String sql = SqlBuilder.buildBatchInsertSql(SegmentTable.TABLE, source.keySet());
......
...@@ -34,5 +34,6 @@ public class SegmentH2TableDefine extends H2TableDefine { ...@@ -34,5 +34,6 @@ public class SegmentH2TableDefine extends H2TableDefine {
@Override public void initialize() { @Override public void initialize() {
addColumn(new H2ColumnDefine(SegmentTable.COLUMN_ID, H2ColumnDefine.Type.Varchar.name())); addColumn(new H2ColumnDefine(SegmentTable.COLUMN_ID, H2ColumnDefine.Type.Varchar.name()));
addColumn(new H2ColumnDefine(SegmentTable.COLUMN_DATA_BINARY, H2ColumnDefine.Type.BINARY.name())); addColumn(new H2ColumnDefine(SegmentTable.COLUMN_DATA_BINARY, H2ColumnDefine.Type.BINARY.name()));
addColumn(new H2ColumnDefine(SegmentTable.COLUMN_TIME_BUCKET, H2ColumnDefine.Type.Bigint.name()));
} }
} }
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册