diff --git a/flink/Dockerfile b/flink/Dockerfile deleted file mode 100644 index cf9e0cdbe9b4..000000000000 --- a/flink/Dockerfile +++ /dev/null @@ -1,16 +0,0 @@ -FROM flink:1.18-java17 - -RUN apt-get update \ - && apt-get install --no-install-recommends -y zip \ - && apt-get clean \ - && rm -rf /var/lib/apt/lists/* \ - && mkdir -p /build - -COPY target/pinpoint-flink-job*.jar /build/pinpoint-flink-job.jar -COPY dockerfiles/build/ /build/ -COPY dockerfiles/docker-bin/ /docker-bin/ -COPY dockerfiles/flink-conf.yml /opt/flink/conf/flink-conf.yaml - -RUN chmod a+x /docker-bin/docker-entrypoint.sh - -ENTRYPOINT ["/docker-bin/docker-entrypoint.sh"] \ No newline at end of file diff --git a/flink/dockerfiles/build/profiles/release/hbase.properties b/flink/dockerfiles/build/profiles/release/hbase.properties deleted file mode 100644 index d036a994cb96..000000000000 --- a/flink/dockerfiles/build/profiles/release/hbase.properties +++ /dev/null @@ -1,29 +0,0 @@ -hbase.client.host=#PINPOINT_ZOOKEEPER_ADDRESS# -hbase.client.port=2181 - -# hbase default:/hbase -hbase.client.znode=/hbase - -# hbase namespace to use default:default -hbase.namespace=default - -# hbase timeout option================================================================================== -# hbase default:true -hbase.client.properties.hbase.ipc.client.tcpnodelay=true -# hbase default:60000 -hbase.client.properties.hbase.rpc.timeout=10000 -# hbase default:Integer.MAX_VALUE -hbase.client.properties.hbase.client.operation.timeout=10000 -# hbase default: 20m -hbase.client.properties.hbase.client.meta.operation.timeout=10000 - -# hbase socket read timeout. default: 200000 -hbase.client.properties.hbase.ipc.client.socket.timeout.read=20000 -# socket write timeout. hbase default: 600000 -hbase.client.properties.hbase.ipc.client.socket.timeout.write=30000 - -#================================================================================== -# hbase client thread pool option -hbase.client.executor.corePoolSize=128 -hbase.client.executor.maxPoolSize=128 -hbase.client.executor.queueCapacity=5120 diff --git a/flink/dockerfiles/build/profiles/release/pinpoint-flink.properties b/flink/dockerfiles/build/profiles/release/pinpoint-flink.properties deleted file mode 100644 index 7d7a6367eee1..000000000000 --- a/flink/dockerfiles/build/profiles/release/pinpoint-flink.properties +++ /dev/null @@ -1,43 +0,0 @@ -# configure l4 ip address to ignore health check logs -collector.l4.ip= - -# base data receiver config --------------------------------------------------------------------- -collector.receiver.base.ip=0.0.0.0 -collector.receiver.base.port=19994 - -# number of tcp worker threads -flink.receiver.base.worker.corePoolSize=8 -flink.receiver.base.worker.maxPoolSize=8 -# capacity of tcp worker queue -flink.receiver.base.worker.queueCapacity=1024 -# monitoring for tcp worker -flink.receiver.base.worker.monitor=true - -# change OS level read/write socket buffer size (for linux) -#sudo sysctl -w net.core.rmem_max= -#sudo sysctl -w net.core.wmem_max= -# check current values using: -#$ /sbin/sysctl -a | grep -e rmem -e wmem - -# number of agent event worker threads -collector.agentEventWorker.corePoolSize=4 -collector.agentEventWorker.maxPoolSize=4 -# capacity of agent event worker queue -collector.agentEventWorker.queueCapacity=1024 - -#collector.admin.password= -#collector.admin.api.rest.active= -#collector.admin.api.jmx.active= - -collector.spanEvent.sequence.limit=10000 - -# flink cluster -flink.cluster.enable=true -flink.cluster.zookeeper.address=#PINPOINT_ZOOKEEPER_ADDRESS# -flink.cluster.zookeeper.znode_root=/pinpoint-cluster -flink.cluster.zookeeper.sessiontimeout=3000 -flink.cluster.zookeeper.retry.interval=5000 -flink.cluster.tcp.port=19994 - -# flink env init -flink.StreamExecutionEnvironment=server \ No newline at end of file diff --git a/flink/dockerfiles/docker-bin/docker-entrypoint.sh b/flink/dockerfiles/docker-bin/docker-entrypoint.sh deleted file mode 100644 index 2d0aeb6a9e60..000000000000 --- a/flink/dockerfiles/docker-bin/docker-entrypoint.sh +++ /dev/null @@ -1,26 +0,0 @@ -#!/bin/bash - -# In flink version < 1.11 we can't use "job cluster" because it's not yet available -if [ "$1" = "standalone-job" ]; then - if [ -e /build/pinpoint-flink-job.jar ]; then - cd /build - zip -d pinpoint-flink-job.jar "profiles/release/hbase.properties" - zip -d pinpoint-flink-job.jar "profiles/release/pinpoint-flink.properties" - sed -i -e "s/#PINPOINT_ZOOKEEPER_ADDRESS#/${PINPOINT_ZOOKEEPER_ADDRESS}/g" profiles/release/* - zip -r pinpoint-flink-job.jar profiles/ - cd - - - mv /build/pinpoint-flink-job.jar pinpoint-flink-job.jar - fi - - shift - - /docker-entrypoint.sh jobmanager & - sleep 5 - if ! flink list | grep 'RUNNING'; then - flink run $@ - fi - wait -else - exec /docker-entrypoint.sh $@ -fi \ No newline at end of file diff --git a/flink/dockerfiles/flink-conf.yml b/flink/dockerfiles/flink-conf.yml deleted file mode 100644 index 1b00c1310265..000000000000 --- a/flink/dockerfiles/flink-conf.yml +++ /dev/null @@ -1,311 +0,0 @@ -################################################################################ -# Licensed to the Apache Software Foundation (ASF) under one -# or more contributor license agreements. See the NOTICE file -# distributed with this work for additional information -# regarding copyright ownership. The ASF licenses this file -# to you under the Apache License, Version 2.0 (the -# "License"); you may not use this file except in compliance -# with the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -################################################################################ - -# These parameters are required for Java 17 support. -# They can be safely removed when using Java 8/11. -env.java.opts.all: --add-exports=java.base/sun.net.util=ALL-UNNAMED --add-exports=java.rmi/sun.rmi.registry=ALL-UNNAMED --add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED --add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED --add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED --add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED --add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED --add-exports=java.security.jgss/sun.security.krb5=ALL-UNNAMED --add-opens=java.base/java.lang=ALL-UNNAMED --add-opens=java.base/java.net=ALL-UNNAMED --add-opens=java.base/java.io=ALL-UNNAMED --add-opens=java.base/java.nio=ALL-UNNAMED --add-opens=java.base/sun.nio.ch=ALL-UNNAMED --add-opens=java.base/java.lang.reflect=ALL-UNNAMED --add-opens=java.base/java.text=ALL-UNNAMED --add-opens=java.base/java.time=ALL-UNNAMED --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.util.concurrent.atomic=ALL-UNNAMED --add-opens=java.base/java.util.concurrent.locks=ALL-UNNAMED -Dhbase.namespace=default - -#============================================================================== -# Common -#============================================================================== - -# The external address of the host on which the JobManager runs and can be -# reached by the TaskManagers and any clients which want to connect. This setting -# is only used in Standalone mode and may be overwritten on the JobManager side -# by specifying the --host parameter of the bin/jobmanager.sh executable. -# In high availability mode, if you use the bin/start-cluster.sh script and setup -# the conf/masters file, this will be taken care of automatically. Yarn -# automatically configure the host name based on the hostname of the node where the -# JobManager runs. - -jobmanager.rpc.address: jobmanager - -# The RPC port where the JobManager is reachable. - -jobmanager.rpc.port: 6123 - -# The host interface the JobManager will bind to. By default, this is localhost, and will prevent -# the JobManager from communicating outside the machine/container it is running on. -# On YARN this setting will be ignored if it is set to 'localhost', defaulting to 0.0.0.0. -# On Kubernetes this setting will be ignored, defaulting to 0.0.0.0. -# -# To enable this, set the bind-host address to one that has access to an outside facing network -# interface, such as 0.0.0.0. - -jobmanager.bind-host: 0.0.0.0 - - -# The total process memory size for the JobManager. -# -# Note this accounts for all memory usage within the JobManager process, including JVM metaspace and other overhead. - -jobmanager.memory.process.size: 1600m - -# The host interface the TaskManager will bind to. By default, this is localhost, and will prevent -# the TaskManager from communicating outside the machine/container it is running on. -# On YARN this setting will be ignored if it is set to 'localhost', defaulting to 0.0.0.0. -# On Kubernetes this setting will be ignored, defaulting to 0.0.0.0. -# -# To enable this, set the bind-host address to one that has access to an outside facing network -# interface, such as 0.0.0.0. - -taskmanager.bind-host: 0.0.0.0 - -# The address of the host on which the TaskManager runs and can be reached by the JobManager and -# other TaskManagers. If not specified, the TaskManager will try different strategies to identify -# the address. -# -# Note this address needs to be reachable by the JobManager and forward traffic to one of -# the interfaces the TaskManager is bound to (see 'taskmanager.bind-host'). -# -# Note also that unless all TaskManagers are running on the same machine, this address needs to be -# configured separately for each TaskManager. - - -# The total process memory size for the TaskManager. -# -# Note this accounts for all memory usage within the TaskManager process, including JVM metaspace and other overhead. - -taskmanager.memory.process.size: 1728m - -# To exclude JVM metaspace and overhead, please, use total Flink memory size instead of 'taskmanager.memory.process.size'. -# It is not recommended to set both 'taskmanager.memory.process.size' and Flink memory. -# -# taskmanager.memory.flink.size: 1280m - -# The number of task slots that each TaskManager offers. Each slot runs one parallel pipeline. - -taskmanager.numberOfTaskSlots: 1 - -# The parallelism used for programs that did not specify and other parallelism. - -parallelism.default: 1 - -# The default file system scheme and authority. -# -# By default file paths without scheme are interpreted relative to the local -# root file system 'file:///'. Use this to override the default and interpret -# relative paths relative to a different file system, -# for example 'hdfs://mynamenode:12345' -# -# fs.default-scheme - -#============================================================================== -# High Availability -#============================================================================== - -# The high-availability mode. Possible options are 'NONE' or 'zookeeper'. -# -# high-availability.type: zookeeper - -# The path where metadata for master recovery is persisted. While ZooKeeper stores -# the small ground truth for checkpoint and leader election, this location stores -# the larger objects, like persisted dataflow graphs. -# -# Must be a durable file system that is accessible from all nodes -# (like HDFS, S3, Ceph, nfs, ...) -# -# high-availability.storageDir: hdfs:///flink/ha/ -# The list of ZooKeeper quorum peers that coordinate the high-availability -# setup. This must be a list of the form: -# "host1:clientPort,host2:clientPort,..." (default clientPort: 2181) -# -# high-availability.zookeeper.quorum: localhost:2181 - - -# ACL options are based on https://zookeeper.apache.org/doc/r3.1.2/zookeeperProgrammers.html#sc_BuiltinACLSchemes -# It can be either "creator" (ZOO_CREATE_ALL_ACL) or "open" (ZOO_OPEN_ACL_UNSAFE) -# The default value is "open" and it can be changed to "creator" if ZK security is enabled -# -# high-availability.zookeeper.client.acl: open - -#============================================================================== -# Fault tolerance and checkpointing -#============================================================================== - -# The backend that will be used to store operator state checkpoints if -# checkpointing is enabled. Checkpointing is enabled when execution.checkpointing.interval > 0. -# -# Execution checkpointing related parameters. Please refer to CheckpointConfig and ExecutionCheckpointingOptions for more details. -# -# execution.checkpointing.interval: 3min -# execution.checkpointing.externalized-checkpoint-retention: [DELETE_ON_CANCELLATION, RETAIN_ON_CANCELLATION] -# execution.checkpointing.max-concurrent-checkpoints: 1 -# execution.checkpointing.min-pause: 0 -# execution.checkpointing.mode: [EXACTLY_ONCE, AT_LEAST_ONCE] -# execution.checkpointing.timeout: 10min -# execution.checkpointing.tolerable-failed-checkpoints: 0 -# execution.checkpointing.unaligned: false -# -# Supported backends are 'hashmap', 'rocksdb', or the -# . -# -# state.backend.type: hashmap - -# Directory for checkpoints filesystem, when using any of the default bundled -# state backends. -# -# state.checkpoints.dir: hdfs://namenode-host:port/flink-checkpoints - -# Default target directory for savepoints, optional. -# -# state.savepoints.dir: hdfs://namenode-host:port/flink-savepoints - -# Flag to enable/disable incremental checkpoints for backends that -# support incremental checkpoints (like the RocksDB state backend). -# -# state.backend.incremental: false - -# The failover strategy, i.e., how the job computation recovers from task failures. -# Only restart tasks that may have been affected by the task failure, which typically includes -# downstream tasks and potentially upstream tasks if their produced data is no longer available for consumption. - -jobmanager.execution.failover-strategy: region - -#============================================================================== -# Rest & web frontend -#============================================================================== - -# The port to which the REST client connects to. If rest.bind-port has -# not been specified, then the server will bind to this port as well. -# -#rest.port: 8081 - -# The address to which the REST client will connect to -# -rest.address: 0.0.0.0 - -# Port range for the REST and web server to bind to. -# -#rest.bind-port: 8080-8090 - -# The address that the REST & web server binds to -# By default, this is localhost, which prevents the REST & web server from -# being able to communicate outside of the machine/container it is running on. -# -# To enable this, set the bind address to one that has access to outside-facing -# network interface, such as 0.0.0.0. -# -rest.bind-address: 0.0.0.0 - -# Flag to specify whether job submission is enabled from the web-based -# runtime monitor. Uncomment to disable. - -#web.submit.enable: false - -# Flag to specify whether job cancellation is enabled from the web-based -# runtime monitor. Uncomment to disable. - -#web.cancel.enable: false - -#============================================================================== -# Advanced -#============================================================================== - -# Override the directories for temporary files. If not specified, the -# system-specific Java temporary directory (java.io.tmpdir property) is taken. -# -# For framework setups on Yarn, Flink will automatically pick up the -# containers' temp directories without any need for configuration. -# -# Add a delimited list for multiple directories, using the system directory -# delimiter (colon ':' on unix) or a comma, e.g.: -# /data1/tmp:/data2/tmp:/data3/tmp -# -# Note: Each directory entry is read from and written to by a different I/O -# thread. You can include the same directory multiple times in order to create -# multiple I/O threads against that directory. This is for example relevant for -# high-throughput RAIDs. -# -# io.tmp.dirs: /tmp - -# The classloading resolve order. Possible values are 'child-first' (Flink's default) -# and 'parent-first' (Java's default). -# -# Child first classloading allows users to use different dependency/library -# versions in their application than those in the classpath. Switching back -# to 'parent-first' may help with debugging dependency issues. -# -# classloader.resolve-order: child-first - -# The amount of memory going to the network stack. These numbers usually need -# no tuning. Adjusting them may be necessary in case of an "Insufficient number -# of network buffers" error. The default min is 64MB, the default max is 1GB. -# -# taskmanager.memory.network.fraction: 0.1 -# taskmanager.memory.network.min: 64mb -# taskmanager.memory.network.max: 1gb - -#============================================================================== -# Flink Cluster Security Configuration -#============================================================================== - -# Kerberos authentication for various components - Hadoop, ZooKeeper, and connectors - -# may be enabled in four steps: -# 1. configure the local krb5.conf file -# 2. provide Kerberos credentials (either a keytab or a ticket cache w/ kinit) -# 3. make the credentials available to various JAAS login contexts -# 4. configure the connector to use JAAS/SASL - -# The below configure how Kerberos credentials are provided. A keytab will be used instead of -# a ticket cache if the keytab path and principal are set. - -# security.kerberos.login.use-ticket-cache: true -# security.kerberos.login.keytab: /path/to/kerberos/keytab -# security.kerberos.login.principal: flink-user - -# The configuration below defines which JAAS login contexts - -# security.kerberos.login.contexts: Client,KafkaClient - -#============================================================================== -# ZK Security Configuration -#============================================================================== - -# Below configurations are applicable if ZK ensemble is configured for security - -# Override below configuration to provide custom ZK service name if configured -# zookeeper.sasl.service-name: zookeeper - -# The configuration below must match one of the values set in "security.kerberos.login.contexts" -# zookeeper.sasl.login-context-name: Client - -#============================================================================== -# HistoryServer -#============================================================================== - -# The HistoryServer is started and stopped via bin/historyserver.sh (start|stop) - -# Directory to upload completed jobs to. Add this directory to the list of -# monitored directories of the HistoryServer as well (see below). -#jobmanager.archive.fs.dir: hdfs:///completed-jobs/ - -# The address under which the web-based HistoryServer listens. -#historyserver.web.address: 0.0.0.0 - -# The port under which the web-based HistoryServer listens. -#historyserver.web.port: 8082 - -# Comma separated list of directories to monitor for completed jobs. -#historyserver.archive.fs.dir: hdfs:///completed-jobs/ - -# Interval in milliseconds for refreshing the monitored directories. -#historyserver.archive.fs.refresh-interval: 10000 - -blob.server.port: 6124 -query.server.port: 6125 \ No newline at end of file diff --git a/flink/pom.xml b/flink/pom.xml deleted file mode 100644 index 1c80378e844d..000000000000 --- a/flink/pom.xml +++ /dev/null @@ -1,47 +0,0 @@ - - - - - - pinpoint - com.navercorp.pinpoint - 3.1.0-SNAPSHOT - - 4.0.0 - - - pinpoint-flink - pinpoint-flink - - - 17 - ${env.JAVA_17_HOME} - - - - - - - - - - - - - \ No newline at end of file diff --git a/pom.xml b/pom.xml index 66ece0295e88..d85dd323823b 100644 --- a/pom.xml +++ b/pom.xml @@ -102,7 +102,6 @@ webhook web-frontend hbase - flink batch pinot @@ -588,11 +587,6 @@ pom ${project.version} - - com.navercorp.pinpoint - pinpoint-flink - ${project.version} - org.springframework