Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
apache
DolphinScheduler
提交
3b562f55
DolphinScheduler
项目概览
apache
/
DolphinScheduler
上一次同步 1 年多
通知
705
Star
9572
Fork
3514
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
DolphinScheduler
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
3b562f55
编写于
2月 26, 2021
作者:
C
chengshiwen
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[Fix-4886][Docker] Fix invalid volume specification in Windows
上级
fbe0edac
变更
7
隐藏空白更改
内联
并排
Showing
7 changed file
with
45 addition
and
53 deletion
+45
-53
docker/build/Dockerfile
docker/build/Dockerfile
+1
-2
docker/build/conf/dolphinscheduler/common.properties.tpl
docker/build/conf/dolphinscheduler/common.properties.tpl
+1
-1
docker/build/conf/dolphinscheduler/env/dolphinscheduler_env.sh.tpl
...ild/conf/dolphinscheduler/env/dolphinscheduler_env.sh.tpl
+9
-9
docker/build/startup-init-conf.sh
docker/build/startup-init-conf.sh
+15
-4
docker/docker-swarm/docker-compose.yml
docker/docker-swarm/docker-compose.yml
+9
-1
docker/docker-swarm/docker-stack.yml
docker/docker-swarm/docker-stack.yml
+10
-8
docker/docker-swarm/dolphinscheduler_env.sh
docker/docker-swarm/dolphinscheduler_env.sh
+0
-28
未找到文件。
docker/build/Dockerfile
浏览文件 @
3b562f55
...
...
@@ -44,11 +44,10 @@ COPY ./startup-init-conf.sh /root/startup-init-conf.sh
COPY
./startup.sh /root/startup.sh
COPY
./conf/dolphinscheduler/*.tpl /opt/dolphinscheduler/conf/
COPY
./conf/dolphinscheduler/logback/* /opt/dolphinscheduler/conf/
COPY
./conf/dolphinscheduler/env/dolphinscheduler_env.sh /opt/dolphinscheduler/conf/env/
COPY
./conf/dolphinscheduler/env/dolphinscheduler_env.sh
.tpl
/opt/dolphinscheduler/conf/env/
RUN
dos2unix /root/checkpoint.sh
&&
\
dos2unix /root/startup-init-conf.sh
&&
\
dos2unix /root/startup.sh
&&
\
dos2unix /opt/dolphinscheduler/conf/env/dolphinscheduler_env.sh
&&
\
dos2unix /opt/dolphinscheduler/script/
*
.sh
&&
\
dos2unix /opt/dolphinscheduler/bin/
*
.sh
&&
\
rm
-rf
/bin/sh
&&
\
...
...
docker/build/conf/dolphinscheduler/common.properties.tpl
浏览文件 @
3b562f55
...
...
@@ -64,7 +64,7 @@ yarn.application.status.address=http://ds1:8088/ws/v1/cluster/apps/%s
yarn.job.history.status.address=http://ds1:19888/ws/v1/history/mapreduce/jobs/%s
# system env path, If you want to set your own path, you need to set this env file to an absolute path
dolphinscheduler.env.path=$
{
DOLPHINSCHEDULER_ENV_PATH
}
#dolphinscheduler.env.path=env/dolphinscheduler_env.sh
development.state=false
# kerberos tgt expire time, unit is hours
...
...
docker/build/conf/dolphinscheduler/env/dolphinscheduler_env.sh
→
docker/build/conf/dolphinscheduler/env/dolphinscheduler_env.sh
.tpl
浏览文件 @
3b562f55
...
...
@@ -15,14 +15,14 @@
# limitations under the License.
#
export
HADOOP_HOME
=
/opt/soft/hadoop
export
HADOOP_CONF_DIR
=
/opt/soft/hadoop/etc/hadoop
export
SPARK_HOME1
=
/opt/soft/spark
1
export
SPARK_HOME2
=
/opt/soft/spark
2
export
PYTHON_HOME
=
/usr/bin/python
export
JAVA_HOME
=
/usr/lib/jvm/java-1.8-openjdk
export
HIVE_HOME
=
/opt/soft/hive
export
FLINK_HOME
=
/opt/soft/flink
export
DATAX_HOME
=
/opt/soft/datax/bin/datax.py
export HADOOP_HOME=
$HADOOP_HOME
export HADOOP_CONF_DIR=
$HADOOP_CONF_DIR
export SPARK_HOME1=
$SPARK_HOME
1
export SPARK_HOME2=
$SPARK_HOME
2
export PYTHON_HOME=
$PYTHON_HOME
export JAVA_HOME=
$JAVA_HOME
export HIVE_HOME=
$HIVE_HOME
export FLINK_HOME=
$FLINK_HOME
export DATAX_HOME=
$DATAX_HOME
export PATH=$HADOOP_HOME/bin:$SPARK_HOME1/bin:$SPARK_HOME2/bin:$PYTHON_HOME:$JAVA_HOME/bin:$HIVE_HOME/bin:$PATH:$FLINK_HOME/bin:$DATAX_HOME:$PATH
docker/build/startup-init-conf.sh
浏览文件 @
3b562f55
...
...
@@ -37,7 +37,17 @@ export DATABASE_PARAMS=${DATABASE_PARAMS:-"characterEncoding=utf8"}
#============================================================================
# Common
#============================================================================
export
DOLPHINSCHEDULER_ENV_PATH
=
${
DOLPHINSCHEDULER_ENV_PATH
:-
"/opt/dolphinscheduler/conf/env/dolphinscheduler_env.sh"
}
# dolphinscheduler env
export
HADOOP_HOME
=
${
HADOOP_HOME
:-
"/opt/soft/hadoop"
}
export
HADOOP_CONF_DIR
=
${
HADOOP_CONF_DIR
:-
"/opt/soft/hadoop/etc/hadoop"
}
export
SPARK_HOME1
=
${
SPARK_HOME1
:-
"/opt/soft/spark1"
}
export
SPARK_HOME2
=
${
SPARK_HOME2
:-
"/opt/soft/spark2"
}
export
PYTHON_HOME
=
${
PYTHON_HOME
:-
"/usr/bin/python"
}
export
JAVA_HOME
=
${
JAVA_HOME
:-
"/usr/lib/jvm/java-1.8-openjdk"
}
export
HIVE_HOME
=
${
HIVE_HOME
:-
"/opt/soft/hive"
}
export
FLINK_HOME
=
${
FLINK_HOME
:-
"/opt/soft/flink"
}
export
DATAX_HOME
=
${
DATAX_HOME
:-
"/opt/soft/datax/bin/datax.py"
}
# common env
export
DOLPHINSCHEDULER_DATA_BASEDIR_PATH
=
${
DOLPHINSCHEDULER_DATA_BASEDIR_PATH
:-
"/tmp/dolphinscheduler"
}
export
DOLPHINSCHEDULER_OPTS
=
${
DOLPHINSCHEDULER_OPTS
:-
""
}
export
RESOURCE_STORAGE_TYPE
=
${
RESOURCE_STORAGE_TYPE
:-
"HDFS"
}
...
...
@@ -83,9 +93,10 @@ export ALERT_LISTEN_HOST=${ALERT_LISTEN_HOST:-"127.0.0.1"}
export
ALERT_PLUGIN_DIR
=
${
ALERT_PLUGIN_DIR
:-
"lib/plugin/alert"
}
echo
"generate app config"
ls
${
DOLPHINSCHEDULER_HOME
}
/conf/ |
grep
".tpl"
|
while
read
lin
e
;
do
find
${
DOLPHINSCHEDULER_HOME
}
/conf/
-name
"*.tpl"
|
while
read
fil
e
;
do
eval
"cat << EOF
$(
cat
${
DOLPHINSCHEDULER_HOME
}
/conf/
${
lin
e
}
)
$(
cat
${
fil
e
}
)
EOF
"
>
${
DOLPHINSCHEDULER_HOME
}
/conf/
${
lin
e
%.*
}
"
>
${
fil
e
%.*
}
done
find
${
DOLPHINSCHEDULER_HOME
}
/conf/
-name
"*.sh"
-exec
chmod
+x
{}
\;
docker/docker-swarm/docker-compose.yml
浏览文件 @
3b562f55
...
...
@@ -162,6 +162,15 @@ services:
WORKER_RESERVED_MEMORY
:
"
0.1"
WORKER_GROUPS
:
"
default"
WORKER_WEIGHT
:
"
100"
HADOOP_HOME
:
"
/opt/soft/hadoop"
HADOOP_CONF_DIR
:
"
/opt/soft/hadoop/etc/hadoop"
SPARK_HOME1
:
"
/opt/soft/spark1"
SPARK_HOME2
:
"
/opt/soft/spark2"
PYTHON_HOME
:
"
/usr/bin/python"
JAVA_HOME
:
"
/usr/lib/jvm/java-1.8-openjdk"
HIVE_HOME
:
"
/opt/soft/hive"
FLINK_HOME
:
"
/opt/soft/flink"
DATAX_HOME
:
"
/opt/soft/datax/bin/datax.py"
DOLPHINSCHEDULER_DATA_BASEDIR_PATH
:
/tmp/dolphinscheduler
ALERT_LISTEN_HOST
:
dolphinscheduler-alert
DATABASE_HOST
:
dolphinscheduler-postgresql
...
...
@@ -183,7 +192,6 @@ services:
-
dolphinscheduler-postgresql
-
dolphinscheduler-zookeeper
volumes
:
-
./dolphinscheduler_env.sh:/opt/dolphinscheduler/conf/env/dolphinscheduler_env.sh
-
dolphinscheduler-worker-data:/tmp/dolphinscheduler
-
dolphinscheduler-logs:/opt/dolphinscheduler/logs
-
dolphinscheduler-resource-local:/dolphinscheduler
...
...
docker/docker-swarm/docker-stack.yml
浏览文件 @
3b562f55
...
...
@@ -156,6 +156,15 @@ services:
WORKER_RESERVED_MEMORY
:
"
0.1"
WORKER_GROUPS
:
"
default"
WORKER_WEIGHT
:
"
100"
HADOOP_HOME
:
"
/opt/soft/hadoop"
HADOOP_CONF_DIR
:
"
/opt/soft/hadoop/etc/hadoop"
SPARK_HOME1
:
"
/opt/soft/spark1"
SPARK_HOME2
:
"
/opt/soft/spark2"
PYTHON_HOME
:
"
/usr/bin/python"
JAVA_HOME
:
"
/usr/lib/jvm/java-1.8-openjdk"
HIVE_HOME
:
"
/opt/soft/hive"
FLINK_HOME
:
"
/opt/soft/flink"
DATAX_HOME
:
"
/opt/soft/datax/bin/datax.py"
DOLPHINSCHEDULER_DATA_BASEDIR_PATH
:
/tmp/dolphinscheduler
ALERT_LISTEN_HOST
:
dolphinscheduler-alert
DATABASE_HOST
:
dolphinscheduler-postgresql
...
...
@@ -173,9 +182,6 @@ services:
timeout
:
5s
retries
:
3
start_period
:
30s
configs
:
-
source
:
dolphinscheduler-worker-task-env
target
:
/opt/dolphinscheduler/conf/env/dolphinscheduler_env.sh
volumes
:
-
dolphinscheduler-worker-data:/tmp/dolphinscheduler
-
dolphinscheduler-logs:/opt/dolphinscheduler/logs
...
...
@@ -193,8 +199,4 @@ volumes:
dolphinscheduler-postgresql
:
dolphinscheduler-zookeeper
:
dolphinscheduler-worker-data
:
dolphinscheduler-logs
:
configs
:
dolphinscheduler-worker-task-env
:
file
:
./dolphinscheduler_env.sh
\ No newline at end of file
dolphinscheduler-logs
:
\ No newline at end of file
docker/docker-swarm/dolphinscheduler_env.sh
已删除
100755 → 0
浏览文件 @
fbe0edac
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
export
HADOOP_HOME
=
/opt/soft/hadoop
export
HADOOP_CONF_DIR
=
/opt/soft/hadoop/etc/hadoop
export
SPARK_HOME1
=
/opt/soft/spark1
export
SPARK_HOME2
=
/opt/soft/spark2
export
PYTHON_HOME
=
/usr/bin/python
export
JAVA_HOME
=
/usr/lib/jvm/java-1.8-openjdk
export
HIVE_HOME
=
/opt/soft/hive
export
FLINK_HOME
=
/opt/soft/flink
export
DATAX_HOME
=
/opt/soft/datax/bin/datax.py
export
PATH
=
$HADOOP_HOME
/bin:
$SPARK_HOME1
/bin:
$SPARK_HOME2
/bin:
$PYTHON_HOME
:
$JAVA_HOME
/bin:
$HIVE_HOME
/bin:
$PATH
:
$FLINK_HOME
/bin:
$DATAX_HOME
:
$PATH
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录