DataSphere安装配置

victorzhzh 2020-05-16

linkis:

links/conf/conf.sh

deployUser=hadoop

### The install home path of Linkis
LINKIS_INSTALL_HOME=/home/hadoop/Install/Linkis #Must provided

WORKSPACE_USER_ROOT_PATH=file:///tmp/linkis/ ##file:// required
### User‘s root hdfs path
HDFS_USER_ROOT_PATH=hdfs:///tmp/linkis ##hdfs:// required

### Path to store job ResultSet:file or hdfs path
RESULT_SET_ROOT_PATH=hdfs:///tmp/linkis

### Provide the DB information of Hive metadata database.
HIVE_META_URL=jdbc:mysql://127.0.0.1:3306/dss?useUnicode=true   #hive元数据,存在mysql里,dss库下
HIVE_META_USER=root
HIVE_META_PASSWORD=root

###HADOOP CONF DIR
HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop

###HIVE CONF DIR
HIVE_CONF_DIR=/usr/local/hive/conf

###SPARK CONF DIR
SPARK_CONF_DIR=/usr/local/spark/conf

linkis/conf/db.sh

### Used to store user‘s custom variables, user‘s configuration, UDFs and functions, while providing the JobHistory service
MYSQL_HOST=127.0.0.1
MYSQL_PORT=3306
MYSQL_DB=dss  #这里选的元数据的路径
MYSQL_USER=root
MYSQL_PASSWORD=root

DSS

dss/conf/config.sh 

deployUser=hadoop

### The install home path of DSS,Must provided
DSS_INSTALL_HOME=/home/hadoop/Install/DSS  #配置DSS的安装路径,配置好之后需要新建这个目录,并且修改这个文件夹权限

### Specifies the user workspace, which is used to store the user‘s script files and log files.
### Generally local directory
WORKSPACE_USER_ROOT_PATH=file:///tmp/linkis/
### Path to store job ResultSet:file or hdfs path
RESULT_SET_ROOT_PATH=hdfs:///tmp/linkis

### 1、DataCheck APPJOINT,This service is used to provide DataCheck capability.
HIVE_META_URL=jdbc:mysql://127.0.0.1:3306/dss?characterEncoding=UTF-8
HIVE_META_USER=root
HIVE_META_PASSWORD=root

#Used to store the azkaban project transformed by DSS
WDS_SCHEDULER_PATH=file:///tmp/wds/scheduler   #在安装目录下提前建好改文件

 DataSphere安装配置

DataSphere安装配置

相关推荐