Skip to content

Commit ccc0aa8

Browse files
authored
Bump tsfile version to 2.1.1 (#94)
* Bump tsfile version to 2.1.1 * fix * fix build * fix build * fix build * fix build * fix build * fix build
1 parent bf06d2e commit ccc0aa8

28 files changed

Lines changed: 733 additions & 624 deletions

File tree

.github/workflows/compile-check.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ jobs:
4848
shell: bash
4949
run: |
5050
if [ "${{ matrix.java }}" -ge 17 ]; then
51-
mvn clean verify -P with-springboot -ntp
51+
mvn clean verify -P with-springboot -P with-all-connectors -P with-examples -ntp
5252
else
53-
mvn clean verify -ntp
53+
mvn clean verify -P with-all-connectors -P with-examples -ntp
5454
fi

connectors/flink-iotdb-connector/src/main/java/org/apache/iotdb/flink/IoTDBSink.java

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -78,14 +78,13 @@ public void open(Configuration parameters) throws Exception {
7878

7979
void initSession() {
8080
if (options.getNodeUrls() != null) {
81-
pool = new SessionPool(
82-
options.getNodeUrls(),
83-
options.getUser(),
84-
options.getPassword(),
85-
sessionPoolSize);
81+
pool =
82+
new SessionPool(
83+
options.getNodeUrls(), options.getUser(), options.getPassword(), sessionPoolSize);
8684
return;
8785
}
88-
pool = new SessionPool(
86+
pool =
87+
new SessionPool(
8988
options.getHost(),
9089
options.getPort(),
9190
options.getUser(),

connectors/flink-iotdb-connector/src/main/java/org/apache/iotdb/flink/options/IoTDBSinkOptions.java

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -43,10 +43,10 @@ public IoTDBSinkOptions(
4343
}
4444

4545
public IoTDBSinkOptions(
46-
List<String> nodeUrls,
47-
String user,
48-
String password,
49-
List<TimeseriesOption> timeseriesOptionList) {
46+
List<String> nodeUrls,
47+
String user,
48+
String password,
49+
List<TimeseriesOption> timeseriesOptionList) {
5050
super(nodeUrls, user, password);
5151
this.timeseriesOptionList = timeseriesOptionList;
5252
}

connectors/flink-tsfile-connector/src/test/java/org/apache/iotdb/flink/util/TSFileConfigUtilCompletenessTest.java

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -92,6 +92,14 @@ public void testTSFileConfigUtilCompleteness() {
9292
"setBooleanEncoding",
9393
"setInt32Encoding",
9494
"setTextEncoding",
95+
"setLz4UseJni",
96+
"setTextCompression",
97+
"setBooleanCompression",
98+
"setInt64Compression",
99+
"setDoubleCompression",
100+
"setInt32Compression",
101+
"setFloatCompression",
102+
"setEncryptKeyFromToken"
95103
};
96104
Set<String> addedSetters = new HashSet<>();
97105
Collections.addAll(addedSetters, setters);

connectors/spark-iotdb-connector/pom.xml

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -53,6 +53,11 @@
5353
<artifactId>iotdb-session</artifactId>
5454
<version>${iotdb.version}</version>
5555
</dependency>
56+
<dependency>
57+
<groupId>org.apache.tsfile</groupId>
58+
<artifactId>tsfile</artifactId>
59+
<version>${tsfile.version}</version>
60+
</dependency>
5661
<dependency>
5762
<groupId>org.apache.tsfile</groupId>
5863
<artifactId>common</artifactId>
@@ -90,6 +95,16 @@
9095
-->
9196
<pluginManagement>
9297
<plugins>
98+
<plugin>
99+
<groupId>org.apache.maven.plugins</groupId>
100+
<artifactId>maven-dependency-plugin</artifactId>
101+
<configuration>
102+
<usedDependencies>
103+
<!-- For some reason the plugin complains if this artifact is included -->
104+
<usedDependency>org.apache.tsfile:common</usedDependency>
105+
</usedDependencies>
106+
</configuration>
107+
</plugin>
93108
<plugin>
94109
<groupId>org.scala-tools</groupId>
95110
<artifactId>maven-scala-plugin</artifactId>

connectors/spark-iotdb-table-connector/spark-iotdb-table-common/src/main/scala/org/apache/iotdb/spark/table/db/IoTDBUtils.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ import org.apache.spark.unsafe.types.UTF8String
2828
import org.apache.tsfile.enums.TSDataType
2929
import org.apache.tsfile.read.common.RowRecord
3030
import org.apache.tsfile.utils.{Binary, DateUtils}
31-
import org.apache.tsfile.write.record.Tablet.ColumnCategory
31+
import org.apache.tsfile.enums.ColumnCategory
3232

3333
import java.util
3434

connectors/spark-iotdb-table-connector/spark-iotdb-table-common/src/main/scala/org/apache/iotdb/spark/table/db/write/IoTDBDataWriter.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ import org.apache.spark.sql.connector.write.{DataWriter, WriterCommitMessage}
2828
import org.apache.spark.sql.types.{NullType, StructType}
2929
import org.apache.tsfile.enums.TSDataType
3030
import org.apache.tsfile.write.record.Tablet
31-
import org.apache.tsfile.write.record.Tablet.ColumnCategory
31+
import org.apache.tsfile.enums.ColumnCategory
3232

3333
class IoTDBDataWriter(options: IoTDBOptions, writeSchema: StructType, tableSchema: StructType) extends DataWriter[InternalRow] with Logging {
3434

examples/iotdb-spring-boot-start/pom.xml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,6 +38,7 @@
3838
<version>3.5.1</version>
3939
<google.java.format.version>1.22.0</google.java.format.version>
4040
<spotless.version>2.43.0</spotless.version>
41+
<iotdb.version>2.0.5</iotdb.version>
4142
</properties>
4243
<dependencies>
4344
<dependency>
@@ -57,7 +58,7 @@
5758
<dependency>
5859
<groupId>org.apache.iotdb</groupId>
5960
<artifactId>isession</artifactId>
60-
<version>2.0.4-SNAPSHOT</version>
61+
<version>${iotdb.version}</version>
6162
</dependency>
6263
</dependencies>
6364
<build>

examples/kafka/src/main/java/org/apache/iotdb/kafka/relational/RelationalConstant.java

Lines changed: 35 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -21,29 +21,39 @@
2121

2222
public class RelationalConstant {
2323

24-
public static final String KAFKA_SERVICE_URL = "172.20.31.71:9094";
25-
public static final String TOPIC = "Kafka-Relational-Test";
26-
public static final String[] IOTDB_URLS = {
27-
"127.0.0.1:6667"
28-
};
29-
public static final String IOTDB_USERNAME = "root";
30-
public static final String IOTDB_PASSWORD = "root";
31-
public static final int SESSION_SIZE = 3;
32-
public static final int CONSUMER_THREAD_NUM = 5;
33-
public static final String[] DATABASES = {"kafka_db1", "kafka_db2"};
34-
public static final String[][] TABLES = {
35-
// database, tableName, columnNames, columnTypes, columnCategories
36-
{"kafka_db1", "tb1", "time,region,model_id,temperature,status", "TIMESTAMP,STRING,STRING,FLOAT,BOOLEAN", "TIME,TAG,ATTRIBUTE,FIELD,FIELD"},
37-
{"kafka_db2", "tb2", "time,plant_id,humidity,status", "TIMESTAMP,STRING,FLOAT,BOOLEAN", "TIME,TAG,FIELD,FIELD"}
38-
};
39-
public static final String[] ALL_DATA = {
40-
// database;tableName;columnName[,columnName]*;value[,value]*[,value[:value]*]*
41-
"kafka_db1;tb1;time,temperature,status;17,3.26,true;18,3.27,false;19,3.28,true",
42-
"kafka_db1;tb1;time,region,model_id,temperature;20,'rgn1','id1',3.31",
43-
"kafka_db2;tb2;time,plant_id,humidity,status;50,'id1',68.7,true",
44-
"kafka_db2;tb2;time,plant_id,humidity,status;51,'id2',68.5,false",
45-
"kafka_db2;tb2;time,plant_id,humidity,status;52,'id3',68.3,true",
46-
"kafka_db2;tb2;time,plant_id,humidity,status;53,'id4',68.8,true",
47-
"kafka_db2;tb2;time,plant_id,humidity,status;54,'id5',68.9,true"
48-
};
24+
public static final String KAFKA_SERVICE_URL = "172.20.31.71:9094";
25+
public static final String TOPIC = "Kafka-Relational-Test";
26+
public static final String[] IOTDB_URLS = {"127.0.0.1:6667"};
27+
public static final String IOTDB_USERNAME = "root";
28+
public static final String IOTDB_PASSWORD = "root";
29+
public static final int SESSION_SIZE = 3;
30+
public static final int CONSUMER_THREAD_NUM = 5;
31+
public static final String[] DATABASES = {"kafka_db1", "kafka_db2"};
32+
public static final String[][] TABLES = {
33+
// database, tableName, columnNames, columnTypes, columnCategories
34+
{
35+
"kafka_db1",
36+
"tb1",
37+
"time,region,model_id,temperature,status",
38+
"TIMESTAMP,STRING,STRING,FLOAT,BOOLEAN",
39+
"TIME,TAG,ATTRIBUTE,FIELD,FIELD"
40+
},
41+
{
42+
"kafka_db2",
43+
"tb2",
44+
"time,plant_id,humidity,status",
45+
"TIMESTAMP,STRING,FLOAT,BOOLEAN",
46+
"TIME,TAG,FIELD,FIELD"
47+
}
48+
};
49+
public static final String[] ALL_DATA = {
50+
// database;tableName;columnName[,columnName]*;value[,value]*[,value[:value]*]*
51+
"kafka_db1;tb1;time,temperature,status;17,3.26,true;18,3.27,false;19,3.28,true",
52+
"kafka_db1;tb1;time,region,model_id,temperature;20,'rgn1','id1',3.31",
53+
"kafka_db2;tb2;time,plant_id,humidity,status;50,'id1',68.7,true",
54+
"kafka_db2;tb2;time,plant_id,humidity,status;51,'id2',68.5,false",
55+
"kafka_db2;tb2;time,plant_id,humidity,status;52,'id3',68.3,true",
56+
"kafka_db2;tb2;time,plant_id,humidity,status;53,'id4',68.8,true",
57+
"kafka_db2;tb2;time,plant_id,humidity,status;54,'id5',68.9,true"
58+
};
4959
}

0 commit comments

Comments
 (0)