Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions .github/workflows/compile-check.yml
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ jobs:
shell: bash
run: |
if [ "${{ matrix.java }}" -ge 17 ]; then
mvn clean verify -P with-springboot -ntp
mvn clean verify -P with-springboot -P with-all-connectors -P with-examples -ntp
else
mvn clean verify -ntp
mvn clean verify -P with-all-connectors -P with-examples -ntp
fi
Original file line number Diff line number Diff line change
Expand Up @@ -78,14 +78,13 @@ public void open(Configuration parameters) throws Exception {

void initSession() {
if (options.getNodeUrls() != null) {
pool = new SessionPool(
options.getNodeUrls(),
options.getUser(),
options.getPassword(),
sessionPoolSize);
pool =
new SessionPool(
options.getNodeUrls(), options.getUser(), options.getPassword(), sessionPoolSize);
return;
}
pool = new SessionPool(
pool =
new SessionPool(
options.getHost(),
options.getPort(),
options.getUser(),
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -43,10 +43,10 @@ public IoTDBSinkOptions(
}

public IoTDBSinkOptions(
List<String> nodeUrls,
String user,
String password,
List<TimeseriesOption> timeseriesOptionList) {
List<String> nodeUrls,
String user,
String password,
List<TimeseriesOption> timeseriesOptionList) {
super(nodeUrls, user, password);
this.timeseriesOptionList = timeseriesOptionList;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -92,6 +92,14 @@ public void testTSFileConfigUtilCompleteness() {
"setBooleanEncoding",
"setInt32Encoding",
"setTextEncoding",
"setLz4UseJni",
"setTextCompression",
"setBooleanCompression",
"setInt64Compression",
"setDoubleCompression",
"setInt32Compression",
"setFloatCompression",
"setEncryptKeyFromToken"
};
Set<String> addedSetters = new HashSet<>();
Collections.addAll(addedSetters, setters);
Expand Down
15 changes: 15 additions & 0 deletions connectors/spark-iotdb-connector/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -53,6 +53,11 @@
<artifactId>iotdb-session</artifactId>
<version>${iotdb.version}</version>
</dependency>
<dependency>
<groupId>org.apache.tsfile</groupId>
<artifactId>tsfile</artifactId>
<version>${tsfile.version}</version>
</dependency>
<dependency>
<groupId>org.apache.tsfile</groupId>
<artifactId>common</artifactId>
Expand Down Expand Up @@ -90,6 +95,16 @@
-->
<pluginManagement>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-dependency-plugin</artifactId>
<configuration>
<usedDependencies>
<!-- For some reason the plugin complains if this artifact is included -->
<usedDependency>org.apache.tsfile:common</usedDependency>
</usedDependencies>
</configuration>
</plugin>
<plugin>
<groupId>org.scala-tools</groupId>
<artifactId>maven-scala-plugin</artifactId>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ import org.apache.spark.unsafe.types.UTF8String
import org.apache.tsfile.enums.TSDataType
import org.apache.tsfile.read.common.RowRecord
import org.apache.tsfile.utils.{Binary, DateUtils}
import org.apache.tsfile.write.record.Tablet.ColumnCategory
import org.apache.tsfile.enums.ColumnCategory

import java.util

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ import org.apache.spark.sql.connector.write.{DataWriter, WriterCommitMessage}
import org.apache.spark.sql.types.{NullType, StructType}
import org.apache.tsfile.enums.TSDataType
import org.apache.tsfile.write.record.Tablet
import org.apache.tsfile.write.record.Tablet.ColumnCategory
import org.apache.tsfile.enums.ColumnCategory

class IoTDBDataWriter(options: IoTDBOptions, writeSchema: StructType, tableSchema: StructType) extends DataWriter[InternalRow] with Logging {

Expand Down
3 changes: 2 additions & 1 deletion examples/iotdb-spring-boot-start/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,7 @@
<version>3.5.1</version>
<google.java.format.version>1.22.0</google.java.format.version>
<spotless.version>2.43.0</spotless.version>
<iotdb.version>2.0.5</iotdb.version>
</properties>
<dependencies>
<dependency>
Expand All @@ -57,7 +58,7 @@
<dependency>
<groupId>org.apache.iotdb</groupId>
<artifactId>isession</artifactId>
<version>2.0.4-SNAPSHOT</version>
<version>${iotdb.version}</version>
</dependency>
</dependencies>
<build>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,29 +21,39 @@

public class RelationalConstant {

public static final String KAFKA_SERVICE_URL = "172.20.31.71:9094";
public static final String TOPIC = "Kafka-Relational-Test";
public static final String[] IOTDB_URLS = {
"127.0.0.1:6667"
};
public static final String IOTDB_USERNAME = "root";
public static final String IOTDB_PASSWORD = "root";
public static final int SESSION_SIZE = 3;
public static final int CONSUMER_THREAD_NUM = 5;
public static final String[] DATABASES = {"kafka_db1", "kafka_db2"};
public static final String[][] TABLES = {
// database, tableName, columnNames, columnTypes, columnCategories
{"kafka_db1", "tb1", "time,region,model_id,temperature,status", "TIMESTAMP,STRING,STRING,FLOAT,BOOLEAN", "TIME,TAG,ATTRIBUTE,FIELD,FIELD"},
{"kafka_db2", "tb2", "time,plant_id,humidity,status", "TIMESTAMP,STRING,FLOAT,BOOLEAN", "TIME,TAG,FIELD,FIELD"}
};
public static final String[] ALL_DATA = {
// database;tableName;columnName[,columnName]*;value[,value]*[,value[:value]*]*
"kafka_db1;tb1;time,temperature,status;17,3.26,true;18,3.27,false;19,3.28,true",
"kafka_db1;tb1;time,region,model_id,temperature;20,'rgn1','id1',3.31",
"kafka_db2;tb2;time,plant_id,humidity,status;50,'id1',68.7,true",
"kafka_db2;tb2;time,plant_id,humidity,status;51,'id2',68.5,false",
"kafka_db2;tb2;time,plant_id,humidity,status;52,'id3',68.3,true",
"kafka_db2;tb2;time,plant_id,humidity,status;53,'id4',68.8,true",
"kafka_db2;tb2;time,plant_id,humidity,status;54,'id5',68.9,true"
};
public static final String KAFKA_SERVICE_URL = "172.20.31.71:9094";
public static final String TOPIC = "Kafka-Relational-Test";
public static final String[] IOTDB_URLS = {"127.0.0.1:6667"};
public static final String IOTDB_USERNAME = "root";
public static final String IOTDB_PASSWORD = "root";
public static final int SESSION_SIZE = 3;
public static final int CONSUMER_THREAD_NUM = 5;
public static final String[] DATABASES = {"kafka_db1", "kafka_db2"};
public static final String[][] TABLES = {
// database, tableName, columnNames, columnTypes, columnCategories
{
"kafka_db1",
"tb1",
"time,region,model_id,temperature,status",
"TIMESTAMP,STRING,STRING,FLOAT,BOOLEAN",
"TIME,TAG,ATTRIBUTE,FIELD,FIELD"
},
{
"kafka_db2",
"tb2",
"time,plant_id,humidity,status",
"TIMESTAMP,STRING,FLOAT,BOOLEAN",
"TIME,TAG,FIELD,FIELD"
}
};
public static final String[] ALL_DATA = {
// database;tableName;columnName[,columnName]*;value[,value]*[,value[:value]*]*
"kafka_db1;tb1;time,temperature,status;17,3.26,true;18,3.27,false;19,3.28,true",
"kafka_db1;tb1;time,region,model_id,temperature;20,'rgn1','id1',3.31",
"kafka_db2;tb2;time,plant_id,humidity,status;50,'id1',68.7,true",
"kafka_db2;tb2;time,plant_id,humidity,status;51,'id2',68.5,false",
"kafka_db2;tb2;time,plant_id,humidity,status;52,'id3',68.3,true",
"kafka_db2;tb2;time,plant_id,humidity,status;53,'id4',68.8,true",
"kafka_db2;tb2;time,plant_id,humidity,status;54,'id5',68.9,true"
};
}
Loading