WO2020258175A1 - Method for automatically installing distributed big data platform cdh - Google Patents
Method for automatically installing distributed big data platform cdh Download PDFInfo
- Publication number
- WO2020258175A1 WO2020258175A1 PCT/CN2019/093339 CN2019093339W WO2020258175A1 WO 2020258175 A1 WO2020258175 A1 WO 2020258175A1 CN 2019093339 W CN2019093339 W CN 2019093339W WO 2020258175 A1 WO2020258175 A1 WO 2020258175A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- configuration
- cdh
- nodes
- service
- node
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/60—Software deployment
- G06F8/61—Installation
Definitions
- the invention relates to the technical field of computer big data, in particular, is a method for automatically installing a distributed big data platform CDH.
- the number of clusters ranges from 3 to hundreds or thousands.
- the more data the higher the probability of manual error and the higher the labor cost. It is urgent to solve the time cost and misoperation problems caused by manual operation.
- the purpose of the present invention is to provide a method for automatically installing a distributed big data platform CDH, which effectively reduces the probability of errors, and reduces labor costs and time costs.
- a CDH method for automated installation of a distributed big data platform specifically includes the following steps:
- S10 Environment preparation; specifically refers to: check the basic environment of each node server and prepare installation resources;
- S20 Script configuration; according to the requirements of the configuration file, the basic information of all node servers and user-defined information are configured;
- S30 Installation and deployment; configure the cluster basic environment and big data operating environment according to the running process of the execution script, including basic service configuration, building NTP service, SSH key-free login, java environment, installing CDH management console service, and verifying service status ;
- step S10 the basic environment of each node server is checked in the step S10, which specifically includes:
- Step S11 Determine whether the operating system version of each node server meets the requirements
- Step S12 Check the system architecture of the operating system of each node server to determine whether the operating system is an x86_64 system architecture;
- Step S13 Check the IP network of each node's server operating system to test whether each node is connected;
- Step S14 Check the configuration file, installation package and execution script of each node server.
- step S20 specifically includes:
- Step S21 Check that the IP addresses, SSH ports, passwords of root users, and custom information installed by the user big data platform of all node servers are configured in the deployment file;
- Step S22 Check whether the configuration information of all node servers is correct, and whether the custom information configuration is reasonable;
- step S30 specifically includes stand-alone configuration installation and online configuration installation:
- the stand-alone configuration installation specifically includes the following steps:
- Step S31 All nodes set the IP of each server as static IP according to the corresponding relationship of the script configuration information, and restart the NETWORK service after the setting is completed;
- Step S32 According to the corresponding relationship of the script configuration information, all the nodes set the machine name of each server in a corresponding unified format, and name them in order;
- Step S33 All nodes modify the hosts file according to the corresponding relationship of the configuration file, adding each node IP and corresponding machine name;
- Step S34 Confirm that all nodes can communicate with each other, and there is no packet loss phenomenon, to ensure the normal network
- Step S35 Variable setting of all nodes: specifically refers to: modifying the value of swappiness and the value of transparent_hugepage of all nodes; adjusting configuration of all nodes to self-startup modification;
- Step S36 Turn off the firewall and selinux of all nodes; cancel the boot-up.
- the online configuration installation specifically refers to: configuring the cluster operating environment according to the existing process, including NTP service, SSH key-free login, JAVA environment, installation of CDH management console service, execution When finished, verify each service status operation; specifically include the following steps:
- Step S311 Configure the operating environment of all servers
- Step S312 all node servers trust each other; determine whether all nodes verify mutual trust successfully, and go to each node through SSH to verify whether the configuration is successful;
- Step S313 Check whether all nodes are in the Chinese time zone; if not, change to the Chinese time zone;
- Step S314 All nodes configure the yum source of the system, with the management node as the HTTP server, and all other nodes are configured; verify that the yum source of all nodes is configured correctly;
- Step S315 Install and build NTP service on all nodes, configure the time server of all nodes as the same time server, and verify whether it is successful;
- Step S316 Install mysql service
- Step S317 Install the basic operation service component
- Step S318 start the service component; specifically: start the CDH management console service according to the configuration management node, and add the server to the self-startup; verify the service status according to the check; verify the startup through log check and port check success.
- step S316 specifically includes:
- step S317 specifically includes:
- step S14 specifically refers to:
- the present invention has the following advantages and beneficial effects:
- the present invention can effectively realize the automatic installation of CDH management console
- the present invention can make the process of building a big data cluster more efficient and reduce the error rate.
- a method for automatically installing a distributed big data platform CDH specifically includes the following steps:
- S10 Environment preparation; specifically refers to: inspecting the basic environment of each node server and preparing for installation resources; the node service and basic environment include the system version, system bitness, and network environment of each node server;
- the checking of the basic environment of each node server in the step S10 specifically includes:
- Step S11 Determine whether the operating system version of each node server meets the requirements; the preferred operating system version is Red Hat Enterprise Linux 6/CentOS 6.x series version;
- Step S12 Check the system architecture of each node's server operating system to determine whether the operating system is an x86_64 system architecture; since the big data platform only supports the x86_64 system architecture, use the uname-r command to view this, and check whether the output contains x86_64.
- Step S13 Check the IP network of each node's server operating system to test whether each node is connected; if not, check the server network environment;
- Step S14 Check the configuration files, installation packages and execution scripts of each node server; specifically:
- S20 Script configuration; according to the requirements of the configuration file, the basic information of all node servers and user-defined information are configured;
- S30 Installation and deployment; configure the cluster basic environment and big data operating environment according to the running process of the execution script, including basic service configuration, building NTP service, SSH key-free login, java environment, installing CDH management console service, and verifying service status ;
- step S20 specifically includes:
- Step S21 Check that the IP addresses, SSH ports, passwords of root users, and custom information installed by the user big data platform of all node servers are configured in the deployment file;
- Step S22 Check whether the configuration information of all node servers is correct, and whether the custom information configuration is reasonable;
- the step S30 specifically includes stand-alone configuration installation and online configuration installation:
- the stand-alone configuration installation specifically refers to: configuring the basic environment of all servers, which specifically includes the following steps:
- Step S31 All nodes set the IP of each server to static IP according to the corresponding relationship of the script configuration information, and restart the NETWORK service after the setting is completed; if not, check the server network according to the abnormal situation of individual nodes and restore the network; if yes, then Go to step S32;
- Step S32 According to the corresponding relationship of the script configuration information, all the nodes set the machine name of each server in a corresponding unified format, and name them in order;
- Step S33 All nodes modify the hosts file according to the corresponding relationship of the configuration file, adding each node IP and corresponding machine name;
- Step S34 Confirm that all nodes can be connected to each other, and there is no packet loss, to ensure that the network is normal; if it cannot be connected, check individual server network conditions and troubleshoot hardware problems; if it can be connected, go to step S35;
- Step S35 Variable setting of all nodes: specifically refers to: modify the values of swappiness and transparent_hugepage of all nodes; adjust configuration of all nodes to be modified after booting; specifically refer to:
- Step S36 Turn off the firewall and selinux of all nodes; cancel the boot-up automatically; specifically:
- it also includes:
- the online configuration installation specifically refers to: configuring the cluster operating environment according to the existing process, including NTP service, SSH key-free login, JAVA environment, and installation of CDH management console After the service is executed, verify each service status operation; specifically include the following steps:
- Step S311 Configure the operating environment of all servers
- Step S312 All node servers trust each other; specifically, this means: copy the private key of each node to each server, refer to the following command SSH-copy-id-i root@bigdata-a-001;
- Each server described here includes this machine, and the corresponding relationship between host name and IP needs to be configured in /etc/hosts when copying the private key.
- step S313 If not, check the reason for the abnormal SSH status of individual servers, and troubleshoot permissions; if the configuration is successful, go to step S313;
- Step S313 Check whether all nodes are in the Chinese time zone; if not, change to the Chinese time zone; specifically:
- Step S314 All nodes configure the yum source of the system, with the management node as the HTTP server, and all other nodes are configured; verify that the yum source of all nodes is configured correctly;
- All nodes verify the yum status and check whether the output is normal by using yum search to verify whether the yum source is configured properly;
- Step S315 Install and build NTP service on all nodes, configure the time server of all nodes as the same time server, and verify whether it is successful;
- Step S316 Install mysql service; specifically including
- the mysql configuration file includes the mysql port number, setting ignore case, and character set encoding
- the configuration management node initializes mysql data, creates access users, and authorizes access; verifies the access status of the mysql database; specifically refers to: the management node initialization database includes the creation of the CDH management console metadata database, and the remote login users defined by the user Create remote access users by name and authorize remote access;
- the management node installs the CDH management console service
- the management node configures the metadata database used by the CDH management console service
- Step S317 Install the basic operation service component
- Step S318 start the service component; specifically: start the CDH management console service according to the configuration management node, and add the server to the self-startup; verify the service status according to the check; verify the startup through log check and port check success.
- the management node server starts the CDH management console service
- This embodiment is further optimized on the basis of the above embodiments.
- most servers are LINUX servers, which support shell scripts.
- This embodiment provides an automated installation of distributed big data platform CDH based on shell scripts under LINUX servers.
- Method, "shell script” technology is a procedural script, a computer program and text file, the content is composed of a series of shell commands, and the content is literally translated by UNIX Shell to operate. It is designed as a scripting language, and its operation mode is similar to that of an interpreted language.
- UNIX SHELL plays the role of a command line interpreter. After reading the shell script, it runs the shell commands in sequence, and then outputs the results. Use shell scripts to perform system management, file operations, etc.
- the server has installed the Red Hat Enterprise Linux 6/CentOS6.x operating system, and the number of operating system bits must be x86_64 bits, the system IP network is normal, the firewall has been turned off, and the RPM installation package of the yum source has been uploaded to the management node.
- step S10 “Environment Preparation” needs to complete the following workflow, install the x86_64-bit Red Hat Enterprise Linux 6/CentOS6.x operating system, manually execute the command to turn off the firewall, check whether the network is normal, and ensure that all nodes interact with each other It can be accessed through SSH, and the RPM installation package of yum source is uploaded to the designated directory of the management node server.
- step S20 “configuration script” needs to complete the following workflow, according to the actual situation of all nodes, according to the remarks in the configuration file, fill in the current node IP, root password, port number and big data platform customization
- the new static IP address and machine name that need to be changed are placed in the configuration script file, and the configuration file and execution script are uploaded to the designated directory of the management node.
- step S30 needs to execute a script on the management node, which will obtain the basic information of the current server from the configuration file, verify the basic environment of the system, and prompt for modification if it encounters an abnormality. After the operation and maintenance personnel are adjusted, The script will continue to execute. It will configure static IP and modify the machine name according to the process. After completion, it will install and configure the cluster operating environment according to the existing process, including NTP service, SSH key-free login, JAVA environment, and installation of CDH management console Services, etc., are executed and the status is verified.
- step S40 "result check" it is necessary to log in to the CDH management console to verify its status.
- step S40 only needs to log in to the CDH management console, check whether the service is started, verify whether the status is normal, and the process of this automated installation of the distributed big data platform CDH is completed.
Landscapes
- Engineering & Computer Science (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Stored Programmes (AREA)
Abstract
A method for automatically installing a distributed big data platform CDH, specifically comprising the following steps: S10, environment preparation: specifically, checking the basic environment of each node server, and preparing installation resources; S20, script configuration: completing the configurations of basic information of all node servers and custom information of a user according to configuration file requirements; S30, installation and deployment: configuring a cluster basic environment and a big data operation environment according to the operation process of an execution script, specifically comprising basic service configuration, NTP service building, SSH key-free login, java environment, CDH management console service installation, and service state verification; and S40, result checking: logging in a CDH management console to check the result. The method can effectively implement the automatic installation of the CDH management console, so that the big data cluster building process is more efficient, and the error rate is reduced.
Description
本发明涉及计算机大数据技术领域,具体的说,是一种自动化安装分布式大数据平台CDH方法。The invention relates to the technical field of computer big data, in particular, is a method for automatically installing a distributed big data platform CDH.
随着信息化时代的迅速发展,大数据的使用越来越广泛,但因安装系统环境的差异,产生了安装分布式大数据平台很多复杂的操作,集群数量越多人工成本越高。With the rapid development of the information age, the use of big data has become more and more widespread. However, due to the difference in the installation system environment, many complicated operations have been generated to install a distributed big data platform. The more clusters, the higher the labor cost.
由于分布式大数据平台需要用到可横向扩展的计算机,集群数量从3个到上百,上千都有,数据越多,人工出错的几率就越高,所耗费的人工成本就越高,亟需解决人工操作带来的时间成本与误操作问题。Because the distributed big data platform needs to use horizontally scalable computers, the number of clusters ranges from 3 to hundreds or thousands. The more data, the higher the probability of manual error and the higher the labor cost. It is urgent to solve the time cost and misoperation problems caused by manual operation.
发明内容Summary of the invention
本发明的目的在于提供一种自动化安装分布式大数据平台CDH方法,有效的降低出错的几率,降低人工成本和时间成本。The purpose of the present invention is to provide a method for automatically installing a distributed big data platform CDH, which effectively reduces the probability of errors, and reduces labor costs and time costs.
本发明通过下述技术方案实现:The present invention is realized through the following technical solutions:
一种自动化安装分布式大数据平台CDH方法,具体包括以下步骤:A CDH method for automated installation of a distributed big data platform specifically includes the following steps:
S10:环境准备;具体是指:对各节点服务器的基础环境进行检查,准备好安装资源;S10: Environment preparation; specifically refers to: check the basic environment of each node server and prepare installation resources;
S20:脚本配置;根据配置文件要求,对所有节点服务器的基本信息及用户的自定义信息配置完毕;S20: Script configuration; according to the requirements of the configuration file, the basic information of all node servers and user-defined information are configured;
S30:安装部署;根据执行脚本的运行流程配置集群基础环境及大数据运行环境,具体包括基础服务配置、搭建NTP服务、SSH免密钥登陆、java环境、安装CDH管理控制台服务以及验证服务状态;S30: Installation and deployment; configure the cluster basic environment and big data operating environment according to the running process of the execution script, including basic service configuration, building NTP service, SSH key-free login, java environment, installing CDH management console service, and verifying service status ;
S40:结果检查;登陆CDH管理控制台查看结果。S40: Check the result; log in to the CDH management console to check the result.
进一步地,为了更好的实现本发明,所述步骤S10中对各节点服务器的基础环境进行检查,具体包括:Further, in order to better implement the present invention, the basic environment of each node server is checked in the step S10, which specifically includes:
步骤S11:判断各节点服务器的操作系统版本是否符合要求;Step S11: Determine whether the operating system version of each node server meets the requirements;
若否,则需要重新安装操作系统;If not, you need to reinstall the operating system;
步骤S12:对各节点服务器操作系统的系统架构检查,判断操作系统是否为x86_64的系统架构;Step S12: Check the system architecture of the operating system of each node server to determine whether the operating system is an x86_64 system architecture;
步骤S13:对各节点服务器操作系统的IP网络进行检查,测试各节点之间是否连通;Step S13: Check the IP network of each node's server operating system to test whether each node is connected;
步骤S14:检查各个节点服务器的配置文件、安装包及执行脚本。Step S14: Check the configuration file, installation package and execution script of each node server.
进一步地,为了更好的实现本发明,所述步骤S20具体包括:Further, in order to better implement the present invention, the step S20 specifically includes:
步骤S21:检查所有节点服务器的IP地址、SSH端口、root用户密码、用户大数据平台安装的自定义信息配置到部署文件里;Step S21: Check that the IP addresses, SSH ports, passwords of root users, and custom information installed by the user big data platform of all node servers are configured in the deployment file;
若否,需要针对提示未合理设置的地方修改;If not, you need to modify the place where the prompt is not set reasonably;
步骤S22:检查所有节点服务器的配置信息是否正确,以及自定义信息配置是否合理;Step S22: Check whether the configuration information of all node servers is correct, and whether the custom information configuration is reasonable;
若否,则需要针对异常和不合理信息做出调整。If not, you need to make adjustments to abnormal and unreasonable information.
进一步地,为了更好的实现本发明,所述步骤S30具体包括单机配置安装和联机配置安装:所述单机配置安装具体包括以下步骤:Further, in order to better implement the present invention, the step S30 specifically includes stand-alone configuration installation and online configuration installation: the stand-alone configuration installation specifically includes the following steps:
步骤S31:所有节点按照脚本配置信息的对应关系,设置各服务器的IP为静态IP,设置完成后重启NETWORK服务;Step S31: All nodes set the IP of each server as static IP according to the corresponding relationship of the script configuration information, and restart the NETWORK service after the setting is completed;
步骤S32:所有节点按照脚本配置信息的对应关系,设置各服务器的机器名为相应的统一格式,并按照顺序命名;Step S32: According to the corresponding relationship of the script configuration information, all the nodes set the machine name of each server in a corresponding unified format, and name them in order;
步骤S33:所有节点按照配置文件的对应关系,修改hosts文件,加入各个节点IP以及对应的机器名;Step S33: All nodes modify the hosts file according to the corresponding relationship of the configuration file, adding each node IP and corresponding machine name;
步骤S34:确认相互各个节点都能连通,且没有丢包现象,保证网络的正常;Step S34: Confirm that all nodes can communicate with each other, and there is no packet loss phenomenon, to ensure the normal network;
步骤S35:所有节点的变量设置:具体是指:修改所有节点swappiness的值和transparent_hugepage的值;所有节点调整配置为开机自启动修改;Step S35: Variable setting of all nodes: specifically refers to: modifying the value of swappiness and the value of transparent_hugepage of all nodes; adjusting configuration of all nodes to self-startup modification;
步骤S36:关闭所有节点的防火墙和selinux;取消开机自启动。Step S36: Turn off the firewall and selinux of all nodes; cancel the boot-up.
进一步地,为了更好的实现本发明,所述联机配置安装具体是指:根据既有流程配置集群运行环境,包括NTP服务、SSH免密钥登陆、JAVA环境、安装CDH管理控制台服务,执行完毕,验证各个服务状态操作;具体包括以下步骤:Further, in order to better implement the present invention, the online configuration installation specifically refers to: configuring the cluster operating environment according to the existing process, including NTP service, SSH key-free login, JAVA environment, installation of CDH management console service, execution When finished, verify each service status operation; specifically include the following steps:
步骤S311:配置所有服务器的运行环境;Step S311: Configure the operating environment of all servers;
步骤S312:所有节点服务器互信;判断所有节点验证互信是否成功,通过SSH到各个节点,验证是否配置成功;Step S312: all node servers trust each other; determine whether all nodes verify mutual trust successfully, and go to each node through SSH to verify whether the configuration is successful;
若否,则检查个别服务器SSH状态异常原因,排查权限的问题;If not, check the reasons for the abnormal SSH status of individual servers and troubleshoot permissions;
步骤S313:检查所有节点是否是中国时区;若否,则更改成中国时区;Step S313: Check whether all nodes are in the Chinese time zone; if not, change to the Chinese time zone;
步骤S314:所有节点配置系统的yum源,以管理节点为HTTP服务器,其余所有节点配置;验证所有节点的yum源是否配置正确;Step S314: All nodes configure the yum source of the system, with the management node as the HTTP server, and all other nodes are configured; verify that the yum source of all nodes is configured correctly;
步骤S315:所有节点安装搭建NTP服务,配置所有节点的时间服务器为同一台时间服务器,并验证是否成功;Step S315: Install and build NTP service on all nodes, configure the time server of all nodes as the same time server, and verify whether it is successful;
若否,则检查个别服务器异常原因,排查是否NTP组件安装失败或者未启动的异常;If not, check the cause of the abnormality of the individual server, and check whether the NTP component installation fails or is not started;
步骤S316:安装mysql服务;Step S316: Install mysql service;
步骤S317:安装基础运行服务组件;Step S317: Install the basic operation service component;
步骤S318:启动服务组件;具体是指:根据所述配置管理节点启动CDH管理控制台服务,并把服务器加入开机自启动当中;根据所述检查验证服务状态;通过日志检查和端口检查,验证启动成功。Step S318: start the service component; specifically: start the CDH management console service according to the configuration management node, and add the server to the self-startup; verify the service status according to the check; verify the startup through log check and port check success.
进一步地,为了更好的实现本发明,所述步骤S316具体包括:Further, in order to better implement the present invention, the step S316 specifically includes:
根据所述配置管理节点安装mysql数据库;Install a mysql database according to the configuration management node;
根据所述配置管理节点修改mysql配置文件;Modify the mysql configuration file according to the configuration management node;
根据所述配置管理节点设置mysql服务为开机自启动;Set the mysql service to be self-starting after booting according to the configuration management node;
根据所述配置管理节点初始化mysql数据、创建访问用户、授权访问;Initialize mysql data, create access users, and authorize access according to the configuration management node;
验证mysql数据库的访问状态。Verify the access status of the mysql database.
进一步地,为了更好的实现本发明,所述步骤S317具体包括:Further, in order to better implement the present invention, the step S317 specifically includes:
根据所述配置所有节点安装java运行环境;Install java operating environment on all nodes according to the configuration;
根据所述配置所有节点安装mysql驱动环境;Install the mysql driver environment on all nodes according to the described configuration;
根据所述配置管理节点安装CDH管理控制台服务;Install the CDH management console service according to the configuration management node;
根据所述配置CDH管理控制台服务使用的元数据库。Configure the metadata database used by the CDH management console service according to the described configuration.
进一步地,为了更好的实现本发明,所述步骤S14具体是指:Further, in order to better implement the present invention, the step S14 specifically refers to:
判断所述管理节点的配置文件是否按照当前集群机器情况修改IP用户、密码、修改后的机器名;Determine whether the configuration file of the management node has modified the IP user, password, and modified machine name according to the current cluster machine situation;
判断所述管理节点目录下是否存在所述配置文件及安装源。Determine whether the configuration file and the installation source exist in the management node directory.
本发明与现有技术相比,具有以下优点及有益效果:Compared with the prior art, the present invention has the following advantages and beneficial effects:
(1)本发明能够有效地实现CDH管理控制台的自动化安装;(1) The present invention can effectively realize the automatic installation of CDH management console;
(2)本发明能够使大数据集群搭建的过程更具高效性、降低失误率。(2) The present invention can make the process of building a big data cluster more efficient and reduce the error rate.
下面结合实施例对本发明作进一步地详细说明,但本发明的实施方式不限于此。The present invention will be further described in detail below in conjunction with examples, but the embodiments of the present invention are not limited thereto.
实施例1:Example 1:
本发明通过下述技术方案实现,一种自动化安装分布式大数据平台CDH方法,具体包括以下步骤:The present invention is realized through the following technical solutions. A method for automatically installing a distributed big data platform CDH specifically includes the following steps:
S10:环境准备;具体是指:对各节点服务器的基础环境进行检查,准备好安装资源;所述节点服务且基础环境包括,各个节点服务器的系统版本、系统位数、网络环境;S10: Environment preparation; specifically refers to: inspecting the basic environment of each node server and preparing for installation resources; the node service and basic environment include the system version, system bitness, and network environment of each node server;
所述步骤S10中对各节点服务器的基础环境进行检查,具体包括:The checking of the basic environment of each node server in the step S10 specifically includes:
步骤S11:判断各节点服务器的操作系统版本是否符合要求;优选的所述操作系统版本 为Red Hat Enterprise Linux 6/CentOS6.x系列版本;Step S11: Determine whether the operating system version of each node server meets the requirements; the preferred operating system version is Red Hat Enterprise Linux 6/CentOS 6.x series version;
若否,则需要重新安装操作系统;If not, you need to reinstall the operating system;
步骤S12:对各节点服务器操作系统的系统架构检查,判断操作系统是否为x86_64的系统架构;由于大数据平台只支持x86_64的系统架构,对此使用uname-r命令查看,检查输出是否包含x86_64。Step S12: Check the system architecture of each node's server operating system to determine whether the operating system is an x86_64 system architecture; since the big data platform only supports the x86_64 system architecture, use the uname-r command to view this, and check whether the output contains x86_64.
步骤S13:对各节点服务器操作系统的IP网络进行检查,测试各节点之间是否连通;若否,则需要对服务器网络环境检查;Step S13: Check the IP network of each node's server operating system to test whether each node is connected; if not, check the server network environment;
步骤S14:检查各个节点服务器的配置文件、安装包及执行脚本;具体是指:Step S14: Check the configuration files, installation packages and execution scripts of each node server; specifically:
判断所述管理节点的配置文件是否按照当前集群机器情况修改IP用户、密码、修改后的机器名;Determine whether the configuration file of the management node has modified the IP user, password, and modified machine name according to the current cluster machine situation;
判断所述管理节点目录下是否存在所述配置文件及安装源。Determine whether the configuration file and the installation source exist in the management node directory.
S20:脚本配置;根据配置文件要求,对所有节点服务器的基本信息及用户的自定义信息配置完毕;S20: Script configuration; according to the requirements of the configuration file, the basic information of all node servers and user-defined information are configured;
S30:安装部署;根据执行脚本的运行流程配置集群基础环境及大数据运行环境,具体包括基础服务配置、搭建NTP服务、SSH免密钥登陆、java环境、安装CDH管理控制台服务以及验证服务状态;S30: Installation and deployment; configure the cluster basic environment and big data operating environment according to the running process of the execution script, including basic service configuration, building NTP service, SSH key-free login, java environment, installing CDH management console service, and verifying service status ;
S40:结果检查;登陆CDH管理控制台查看结果。S40: Check the result; log in to the CDH management console to check the result.
本实施例的其他部分与上述实施例相同,故不再赘述。The other parts of this embodiment are the same as the above-mentioned embodiments, so they will not be repeated.
实施例2:Example 2:
本实施例在上述实施例的基础上做进一步优化,所述步骤S20具体包括:This embodiment is further optimized on the basis of the foregoing embodiment, and the step S20 specifically includes:
步骤S21:检查所有节点服务器的IP地址、SSH端口、root用户密码、用户大数据平台安装的自定义信息配置到部署文件里;Step S21: Check that the IP addresses, SSH ports, passwords of root users, and custom information installed by the user big data platform of all node servers are configured in the deployment file;
若否,需要针对提示未合理设置的地方修改;If not, you need to modify the place where the prompt is not set reasonably;
步骤S22:检查所有节点服务器的配置信息是否正确,以及自定义信息配置是否合理;Step S22: Check whether the configuration information of all node servers is correct, and whether the custom information configuration is reasonable;
若否,则需要针对异常和不合理信息做出调整。If not, you need to make adjustments to abnormal and unreasonable information.
本实施例的其他部分与上述实施例相同,故不再赘述。The other parts of this embodiment are the same as the above-mentioned embodiments, so they will not be repeated.
实施例3:Example 3:
本实施例在上述实施例的基础上做进一步优化,所述步骤S30具体包括单机配置安装和联机配置安装:所述单机配置安装具体是指:配置所有服务器的基础环境,具体包括以下步骤:This embodiment is further optimized on the basis of the foregoing embodiment. The step S30 specifically includes stand-alone configuration installation and online configuration installation: the stand-alone configuration installation specifically refers to: configuring the basic environment of all servers, which specifically includes the following steps:
步骤S31:所有节点按照脚本配置信息的对应关系,设置各服务器的IP为静态IP,设置 完成后重启NETWORK服务;若否,根据个别节点的异常情况,检查服务器网络,重新恢复网络;若是,则进入步骤S32;Step S31: All nodes set the IP of each server to static IP according to the corresponding relationship of the script configuration information, and restart the NETWORK service after the setting is completed; if not, check the server network according to the abnormal situation of individual nodes and restore the network; if yes, then Go to step S32;
步骤S32:所有节点按照脚本配置信息的对应关系,设置各服务器的机器名为相应的统一格式,并按照顺序命名;Step S32: According to the corresponding relationship of the script configuration information, all the nodes set the machine name of each server in a corresponding unified format, and name them in order;
步骤S33:所有节点按照配置文件的对应关系,修改hosts文件,加入各个节点IP以及对应的机器名;Step S33: All nodes modify the hosts file according to the corresponding relationship of the configuration file, adding each node IP and corresponding machine name;
步骤S34:确认相互各个节点都能连通,且没有丢包现象,保证网络的正常;若不能连通,则检查个别服务器网络情况,排查硬件的问题;若能连通,则进入步骤S35;Step S34: Confirm that all nodes can be connected to each other, and there is no packet loss, to ensure that the network is normal; if it cannot be connected, check individual server network conditions and troubleshoot hardware problems; if it can be connected, go to step S35;
步骤S35:所有节点的变量设置:具体是指:修改所有节点swappiness的值和transparent_hugepage的值;所有节点调整配置为开机自启动修改;具体是指:Step S35: Variable setting of all nodes: specifically refers to: modify the values of swappiness and transparent_hugepage of all nodes; adjust configuration of all nodes to be modified after booting; specifically refer to:
修改所有节点swappiness的值,设置其在运行时更改该设置,将/proc/sys/vm/swappiness设置为0;Modify the value of swappiness for all nodes, set it to change the setting at runtime, and set /proc/sys/vm/swappiness to 0;
所有节点调整配置其为开机自启动,在添加/etc/sysctl.conf以在重启后保存该设置,该设置重启后不会丢失cat>>/etc/sysctl.conf<<EOF vm.swappiness=0EOF;All nodes are adjusted and configured to start automatically after booting. Add /etc/sysctl.conf to save this setting after restarting. This setting will not be lost after restarting cat>>/etc/sysctl.conf<<EOF vm.swappiness=0EOF ;
所有节点调整配置,修改transparent_hugepage的值,执行立即生效All nodes adjust the configuration, modify the value of transparent_hugepage, the execution takes effect immediately
echo never>/sys/kernel/mm/transparent_hugepage/defrag;echo never>/sys/kernel/mm/transparent_hugepage/defrag;
所有节点调整配置其为开机自启动,加入启动脚本,以永久生效echo"echo never>/sys/kernel/mm/transparent_hugepage/defrag">>/etc/rc.local;All nodes are adjusted and configured to start automatically after booting, and a startup script is added to make it permanent. echo "echo never>/sys/kernel/mm/transparent_hugepage/defrag">>/etc/rc.local;
步骤S36:关闭所有节点的防火墙和selinux;取消开机自启动;具体是指:Step S36: Turn off the firewall and selinux of all nodes; cancel the boot-up automatically; specifically:
关闭所有节点的防火墙,service iptables saveTurn off the firewall of all nodes, service iptables save
service iptables stop;service iptables stop;
chkconfig iptables off;chkconfig iptables off;
service ip6tables save;service ip6tables save;
service ip6tables stop;service ip6tables stop;
chkconfig ip6tables off;chkconfig ip6tables off;
优选的,还包括:Preferably, it also includes:
关闭所有节点的selinux,Close selinux of all nodes,
setenforce 0sed-i.bak's/SELINUX=enforcing/SELINUX=disabled/'/etc/sysconfig/selinux;setenforce 0sed-i.bak's/SELINUX=enforcing/SELINUX=disabled/'/etc/sysconfig/selinux;
sed-i.bak's/SELINUX=enforcing/SELINUX=disabled/'/etc/selinux/config;sed-i.bak's/SELINUX=enforcing/SELINUX=disabled/'/etc/selinux/config;
所有节点执行reboot重启;All nodes perform reboot to restart;
本实施例的其他部分与上述实施例相同,故不再赘述。The other parts of this embodiment are the same as the above-mentioned embodiments, so they will not be repeated.
实施例4:Example 4:
本实施例在上述实施例的基础上做进一步优化,所述联机配置安装具体是指:根据既有流程配置集群运行环境,包括NTP服务、SSH免密钥登陆、JAVA环境、安装CDH管理控制台服务,执行完毕,验证各个服务状态操作;具体包括以下步骤:This embodiment is further optimized on the basis of the above embodiment. The online configuration installation specifically refers to: configuring the cluster operating environment according to the existing process, including NTP service, SSH key-free login, JAVA environment, and installation of CDH management console After the service is executed, verify each service status operation; specifically include the following steps:
步骤S311:配置所有服务器的运行环境;Step S311: Configure the operating environment of all servers;
步骤S312:所有节点服务器互信;具体是指:将各个节点私钥拷贝至每一台服务器,参考命令如下SSH-copy-id-i root@bigdata-a-001;Step S312: All node servers trust each other; specifically, this means: copy the private key of each node to each server, refer to the following command SSH-copy-id-i root@bigdata-a-001;
此处所描述的每一台服务器包括本机,在进行私钥拷贝时需要在/etc/hosts中配置主机名与IP对应关系。Each server described here includes this machine, and the corresponding relationship between host name and IP needs to be configured in /etc/hosts when copying the private key.
判断所有节点验证互信是否成功,通过SSH到各个节点,验证是否配置成功;Judge whether all nodes verify mutual trust successfully, go to each node through SSH, and verify whether the configuration is successful;
若否,则检查个别服务器SSH状态异常原因,排查权限的问题;若配置成功,则进入步骤S313;If not, check the reason for the abnormal SSH status of individual servers, and troubleshoot permissions; if the configuration is successful, go to step S313;
步骤S313:检查所有节点是否是中国时区;若否,则更改成中国时区;具体是指:Step S313: Check whether all nodes are in the Chinese time zone; if not, change to the Chinese time zone; specifically:
检查所有节点是否是中国时区,通过date检查返回结果,是否包含CST;Check whether all nodes are in the Chinese time zone, and check the returned result by date, whether it contains CST;
若否,则更改成中国时区ln-sf/usr/share/zoneinfo/Asia/Shanghai/etc/localtime;If not, change to China time zone ln-sf/usr/share/zoneinfo/Asia/Shanghai/etc/localtime;
步骤S314:所有节点配置系统的yum源,以管理节点为HTTP服务器,其余所有节点配置;验证所有节点的yum源是否配置正确;Step S314: All nodes configure the yum source of the system, with the management node as the HTTP server, and all other nodes are configured; verify that the yum source of all nodes is configured correctly;
所有节点验证yum状态,通过使用yum search检查输出是否正常,以验证yum源是否配置正常;All nodes verify the yum status and check whether the output is normal by using yum search to verify whether the yum source is configured properly;
若否,则排查yum源yum.repos.d文件夹下的配置repo文件是否错误;If not, check whether the configuration repo file in the yum.repos.d folder of the yum source is wrong;
步骤S315:所有节点安装搭建NTP服务,配置所有节点的时间服务器为同一台时间服务器,并验证是否成功;Step S315: Install and build NTP service on all nodes, configure the time server of all nodes as the same time server, and verify whether it is successful;
若否,则检查个别服务器异常原因,排查是否NTP组件安装失败或者未启动的异常;If not, check the cause of the abnormality of the individual server, and check whether the NTP component installation fails or is not started;
步骤S316:安装mysql服务;具体包括Step S316: Install mysql service; specifically including
根据所述配置管理节点安装mysql数据库;Install a mysql database according to the configuration management node;
根据所述配置管理节点修改mysql配置文件;所述mysql配置文件包括mysql端口号、设置忽略大小写、字符集编码;Modify the mysql configuration file according to the configuration management node; the mysql configuration file includes the mysql port number, setting ignore case, and character set encoding;
根据所述配置管理节点设置mysql服务为开机自启动;Set the mysql service to be self-starting after booting according to the configuration management node;
根据所述配置管理节点初始化mysql数据、创建访问用户、授权访问;验证mysql数据 库的访问状态;具体是指:管理节点初始化数据库包含创建CDH管理控制台元数据库,并根据用户自定义的远程登陆用户名创建远程访问用户,授权远程访问;According to the configuration management node initializes mysql data, creates access users, and authorizes access; verifies the access status of the mysql database; specifically refers to: the management node initialization database includes the creation of the CDH management console metadata database, and the remote login users defined by the user Create remote access users by name and authorize remote access;
所有节点验证数据库连接状态;All nodes verify the database connection status;
若否,则检查个别服务器异常原因,排查数据库是否启动,或者用户远程访问授权失败;If not, check the causes of individual server exceptions, check whether the database is started, or the user's remote access authorization fails;
若是,则所有节点安装java环境及安装数据库连接,yum-y install oracle-j2sdk1.7mysql-connector-java;If yes, install java environment and database connection on all nodes, yum-y install oracle-j2sdk1.7mysql-connector-java;
若否,则检查个别服务器异常原因,排查是否存在执行权限问题;If not, check the reason for the abnormality of the individual server and check whether there is an execution authority problem;
若是,则管理节点安装CDH管理控制台服务;If it is, the management node installs the CDH management console service;
若否,则检查管理节点服务器异常原因,排查是否存在缺失依赖的问题;If not, check the reason for the abnormality of the management node server, and check whether there is a problem of missing dependencies;
若是,则管理节点配置CDH管理控制台服务使用的元数据库;If yes, the management node configures the metadata database used by the CDH management console service;
若否,则检查管理节点服务器mysql数据库,检查数据库连接是否存在异常;If not, check the mysql database of the management node server to check whether the database connection is abnormal;
步骤S317:安装基础运行服务组件;Step S317: Install the basic operation service component;
步骤S318:启动服务组件;具体是指:根据所述配置管理节点启动CDH管理控制台服务,并把服务器加入开机自启动当中;根据所述检查验证服务状态;通过日志检查和端口检查,验证启动成功。Step S318: start the service component; specifically: start the CDH management console service according to the configuration management node, and add the server to the self-startup; verify the service status according to the check; verify the startup through log check and port check success.
管理节点服务器启动CDH管理控制台服务;The management node server starts the CDH management console service;
若否,则检查管理节点服务器异常日志,排查是否存在数据库连接异常;If not, check the abnormal log of the management node server to check whether there is an abnormal database connection;
若是,则通过日志检查和端口检查,验证启动成功。If it is, the log check and port check are passed to verify that the startup is successful.
本实施例的其他部分与上述实施例相同,故不再赘述。The other parts of this embodiment are the same as the above-mentioned embodiments, so they will not be repeated.
实施例5:Example 5:
本实施例在上述实施例的基础上做进一步优化,目前大多数服务器都是LINUX服务器,其支持shell脚本,本实施例提供于LINUX服务器下基于shell脚本的一种自动化安装分布式大数据平台CDH方法,“shell脚本”技术是程序化脚本,是一种计算机程序与文本文件,内容由一连串的shell命令组成,经由UNIX SHELL直译其内容后运作。被当成是一种脚本语言来设计,其运作方式与解释型语言相当,由UNIX SHELL扮演命令行解释器的角色,在读取shell脚本之后,依序运行其中的shell命令,之后输出结果。利用shell脚本可以进行系统管理,文件操作等。This embodiment is further optimized on the basis of the above embodiments. At present, most servers are LINUX servers, which support shell scripts. This embodiment provides an automated installation of distributed big data platform CDH based on shell scripts under LINUX servers. Method, "shell script" technology is a procedural script, a computer program and text file, the content is composed of a series of shell commands, and the content is literally translated by UNIX Shell to operate. It is designed as a scripting language, and its operation mode is similar to that of an interpreted language. UNIX SHELL plays the role of a command line interpreter. After reading the shell script, it runs the shell commands in sequence, and then outputs the results. Use shell scripts to perform system management, file operations, etc.
S10,环境准备;S10, environmental preparation;
要求服务器已经安装了Red Hat Enterprise Linux 6/CentOS6.x操作系统,并且操作系统位数必须是x86_64位,系统IP网络正常,防火墙已经关闭,yum源的RPM安装包已经上传到管理节点。It is required that the server has installed the Red Hat Enterprise Linux 6/CentOS6.x operating system, and the number of operating system bits must be x86_64 bits, the system IP network is normal, the firewall has been turned off, and the RPM installation package of the yum source has been uploaded to the management node.
在具体实现过程中,步骤S10“环境准备”需要完成以下的工作流程,安装x86_64位的Red Hat Enterprise Linux 6/CentOS6.x操作系统,手动执行命令关闭防火墙,检查网络是否正常,保证所有节点相互间能够SSH访问,yum源的RPM安装包上传到管理节点服务器指定目录下。In the specific implementation process, step S10 "Environment Preparation" needs to complete the following workflow, install the x86_64-bit Red Hat Enterprise Linux 6/CentOS6.x operating system, manually execute the command to turn off the firewall, check whether the network is normal, and ensure that all nodes interact with each other It can be accessed through SSH, and the RPM installation package of yum source is uploaded to the designated directory of the management node server.
S20,脚本配置;S20, script configuration;
要求根据当前所有节点的情况以及配置文件的备注,完成配置文件的修改。It is required to complete the modification of the configuration file according to the current situation of all nodes and the remarks of the configuration file.
在具体执行过程中,步骤S20“配置脚本”需要完成以下的工作流程,根据所有节点的实际情况,按照配置文件的备注,填写当前各节点的IP、root密码、端口号以及大数据平台自定义需要更改的新的静态IP地址和机器名到配置脚本文件里面,把配置文件和执行脚本上传到管理节点指定目录。In the specific execution process, step S20 "configuration script" needs to complete the following workflow, according to the actual situation of all nodes, according to the remarks in the configuration file, fill in the current node IP, root password, port number and big data platform customization The new static IP address and machine name that need to be changed are placed in the configuration script file, and the configuration file and execution script are uploaded to the designated directory of the management node.
S30,安装部署;S30, installation and deployment;
在具体执行过程中,步骤S30需要在管理节点执行脚本,其会从配置文件里面获取当前服务器的基本信息,验证系统的基础环境,如果遇到异常会提示修改,待运维人员调整完毕之后,脚本会继续执行,其会根据流程配置静态IP和修改机器名等操作,完成之后会根据既有流程安装配置集群运行环境,包括NTP服务、SSH免密钥登陆、JAVA环境、安装CDH管理控制台服务等,执行完毕,验证状态。In the specific execution process, step S30 needs to execute a script on the management node, which will obtain the basic information of the current server from the configuration file, verify the basic environment of the system, and prompt for modification if it encounters an abnormality. After the operation and maintenance personnel are adjusted, The script will continue to execute. It will configure static IP and modify the machine name according to the process. After completion, it will install and configure the cluster operating environment according to the existing process, including NTP service, SSH key-free login, JAVA environment, and installation of CDH management console Services, etc., are executed and the status is verified.
S40,结果检查,登陆CDH管理控制台界面检查状态;S40, check the result, log in to the CDH management console interface to check the status;
本申请实施例中,步骤S40“结果检查”,需要登陆CDH管理控制台,验证其状态。In the embodiment of the present application, in step S40 "result check", it is necessary to log in to the CDH management console to verify its status.
在具体执行过程中,步骤S40只需要登陆CDH管理控制台,查看服务是否启动,验证状态是否正常,本自动化安装分布式大数据平台CDH的流程完毕。In the specific execution process, step S40 only needs to log in to the CDH management console, check whether the service is started, verify whether the status is normal, and the process of this automated installation of the distributed big data platform CDH is completed.
本实施例的其他部分与上述实施例相同,故不再赘述。The other parts of this embodiment are the same as the above-mentioned embodiments, so they will not be repeated.
以上所述,仅是本发明的较佳实施例,并非对本发明做任何形式上的限制,凡是依据本发明的技术实质对以上实施例所作的任何简单修改、等同变化,均落入本发明的保护范围之内。The above are only preferred embodiments of the present invention and do not limit the present invention in any form. Any simple modification or equivalent change made to the above embodiments based on the technical essence of the present invention shall fall into the scope of the present invention. Within the scope of protection.
Claims (8)
- 一种自动化安装分布式大数据平台CDH方法,其特征在于:具体包括以下步骤:A method for automatically installing a distributed big data platform CDH is characterized in that it specifically includes the following steps:S10:环境准备;具体是指:对各节点服务器的基础环境进行检查,准备好安装资源;S10: Environment preparation; specifically refers to: check the basic environment of each node server and prepare installation resources;S20:脚本配置;根据配置文件要求,对所有节点服务器的基本信息及用户的自定义信息配置完毕;S20: Script configuration; according to the requirements of the configuration file, the basic information of all node servers and user-defined information are configured;S30:安装部署;根据执行脚本的运行流程配置集群基础环境及大数据运行环境,具体包括基础服务配置、搭建NTP服务、SSH免密钥登陆、java环境、安装CDH管理控制台服务以及验证服务状态;S30: Installation and deployment; configure the cluster basic environment and big data operating environment according to the running process of the execution script, including basic service configuration, building NTP service, SSH key-free login, java environment, installing CDH management console service, and verifying service status ;S40:结果检查;登陆CDH管理控制台查看结果。S40: Check the result; log in to the CDH management console to check the result.
- 根据权利要求1所述的一种自动化安装分布式大数据平台CDH方法,其特征在于:所述步骤S10中对各节点服务器的基础环境进行检查,具体包括:The method for automatically installing a distributed big data platform CDH according to claim 1, characterized in that: the basic environment of each node server is checked in the step S10, which specifically includes:步骤S11:判断各节点服务器的操作系统版本是否符合要求;Step S11: Determine whether the operating system version of each node server meets the requirements;若否,则需要重新安装操作系统;If not, you need to reinstall the operating system;步骤S12:对各节点服务器操作系统的系统架构检查,判断操作系统是否为x86_64的系统架构;Step S12: Check the system architecture of the operating system of each node server to determine whether the operating system is an x86_64 system architecture;步骤S13:对各节点服务器操作系统的IP网络进行检查,测试各节点之间是否连通;Step S13: Check the IP network of each node's server operating system to test whether each node is connected;步骤S14:检查各个节点服务器的配置文件、安装包及执行脚本。Step S14: Check the configuration file, installation package and execution script of each node server.
- 根据权利要求2所述的一种自动化安装分布式大数据平台CDH方法,其特征在于:所述步骤S20具体包括:The method for automatically installing a distributed big data platform CDH according to claim 2, wherein said step S20 specifically includes:步骤S21:检查所有节点服务器的IP地址、SSH端口、root用户密码、用户大数据平台安装的自定义信息配置到部署文件里;Step S21: Check that the IP addresses, SSH ports, passwords of root users, and custom information installed by the user big data platform of all node servers are configured in the deployment file;若否,需要针对提示未合理设置的地方修改;If not, you need to modify the place where the prompt is not set reasonably;步骤S22:检查所有节点服务器的配置信息是否正确,以及自定义信息配置是否合理;Step S22: Check whether the configuration information of all node servers is correct, and whether the custom information configuration is reasonable;若否,则需要针对异常和不合理信息做出调整。If not, you need to make adjustments to abnormal and unreasonable information.
- 根据权利要求1所述的一种自动化安装分布式大数据平台CDH方法,其特征在于:所述步骤S30具体包括单机配置安装和联机配置安装:所述单机配置安装具体包括以下步骤:The method for automatically installing a distributed big data platform CDH according to claim 1, characterized in that: the step S30 specifically includes a stand-alone configuration installation and an online configuration installation: the stand-alone configuration installation specifically includes the following steps:步骤S31:所有节点按照脚本配置信息的对应关系,设置各服务器的IP为静态IP,设置完成后重启NETWORK服务;Step S31: All nodes set the IP of each server as static IP according to the corresponding relationship of the script configuration information, and restart the NETWORK service after the setting is completed;步骤S32:所有节点按照脚本配置信息的对应关系,设置各服务器的机器名为相应的统一格式,并按照顺序命名;Step S32: According to the corresponding relationship of the script configuration information, all the nodes set the machine name of each server in a corresponding unified format, and name them in order;步骤S33:所有节点按照配置文件的对应关系,修改hosts文件,加入各个节点IP以及对应的机器名;Step S33: All nodes modify the hosts file according to the corresponding relationship of the configuration file, adding each node IP and corresponding machine name;步骤S34:确认相互各个节点都能连通,且没有丢包现象,保证网络的正常;Step S34: Confirm that all nodes can communicate with each other, and there is no packet loss phenomenon, to ensure the normal network;步骤S35:所有节点的变量设置:具体是指:修改所有节点swappiness的值和transparent_hugepage的值;所有节点调整配置为开机自启动修改;Step S35: Variable setting of all nodes: specifically refers to: modifying the value of swappiness and the value of transparent_hugepage of all nodes; adjusting configuration of all nodes to self-startup modification;步骤S36:关闭所有节点的防火墙和selinux;取消开机自启动。Step S36: Turn off the firewall and selinux of all nodes; cancel the boot-up.
- 根据权利要求4所述的一种自动化安装分布式大数据平台CDH方法,其特征在于:所述联机配置安装具体是指:根据既有流程配置集群运行环境,包括NTP服务、SSH免密钥登陆、JAVA环境、安装CDH管理控制台服务,执行完毕,验证各个服务状态操作;具体包括以下步骤:The CDH method for automated installation of a distributed big data platform according to claim 4, characterized in that: the online configuration installation specifically refers to: configuring a cluster operating environment according to an existing process, including NTP service and SSH key-free login , JAVA environment, install CDH management console service, after execution, verify each service status operation; specifically include the following steps:步骤S311:配置所有服务器的运行环境;Step S311: Configure the operating environment of all servers;步骤S312:所有节点服务器互信;判断所有节点验证互信是否成功,通过SSH到各个节点,验证是否配置成功;Step S312: all node servers trust each other; determine whether all nodes verify mutual trust successfully, and go to each node through SSH to verify whether the configuration is successful;若否,则检查个别服务器SSH状态异常原因,排查权限的问题;If not, check the reasons for the abnormal SSH status of individual servers and troubleshoot permissions;步骤S313:检查所有节点是否是中国时区;若否,则更改成中国时区;Step S313: Check whether all nodes are in the Chinese time zone; if not, change to the Chinese time zone;步骤S314:所有节点配置系统的yum源,以管理节点为HTTP服务器,其余所有节点配置;验证所有节点的yum源是否配置正确;Step S314: All nodes configure the yum source of the system, with the management node as the HTTP server, and all other nodes are configured; verify that the yum source of all nodes is configured correctly;步骤S315:所有节点安装搭建NTP服务,配置所有节点的时间服务器为同一台时间服务器,并验证是否成功;Step S315: Install and build NTP service on all nodes, configure the time server of all nodes as the same time server, and verify whether it is successful;若否,则检查个别服务器异常原因,排查是否NTP组件安装失败或者未启动的异常;If not, check the cause of the abnormality of the individual server, and check whether the NTP component installation fails or is not started;步骤S316:安装mysql服务;Step S316: Install mysql service;步骤S317:安装基础运行服务组件;Step S317: Install the basic operation service component;步骤S318:启动服务组件;具体是指:根据所述配置管理节点启动CDH管理控制台服务,并把服务器加入开机自启动当中;根据所述检查验证服务状态;通过日志检查和端口检查,验证启动成功。Step S318: start the service component; specifically: start the CDH management console service according to the configuration management node, and add the server to the self-startup; verify the service status according to the check; verify the startup through log check and port check success.
- 根据权利要求5所述的一种自动化安装分布式大数据平台CDH方法,其特征在于:所述步骤S316具体包括:The method for automatically installing a distributed big data platform CDH according to claim 5, wherein said step S316 specifically includes:根据所述配置管理节点安装mysql数据库;Install a mysql database according to the configuration management node;根据所述配置管理节点修改mysql配置文件;Modify the mysql configuration file according to the configuration management node;根据所述配置管理节点设置mysql服务为开机自启动;Set the mysql service to be self-starting after booting according to the configuration management node;根据所述配置管理节点初始化mysql数据、创建访问用户、授权访问;Initialize mysql data, create access users, and authorize access according to the configuration management node;验证mysql数据库的访问状态。Verify the access status of the mysql database.
- 根据权利要求6所述的一种自动化安装分布式大数据平台CDH方法,其特征在于:所 述步骤S317具体包括:The method for automatically installing a distributed big data platform CDH according to claim 6, wherein said step S317 specifically includes:根据所述配置所有节点安装java运行环境;Install java operating environment on all nodes according to the configuration;根据所述配置所有节点安装mysql驱动环境;Install the mysql driver environment on all nodes according to the described configuration;根据所述配置管理节点安装CDH管理控制台服务;Install the CDH management console service according to the configuration management node;根据所述配置CDH管理控制台服务使用的元数据库。Configure the metadata database used by the CDH management console service according to the described configuration.
- 根据权利要求7所述的一种自动化安装分布式大数据平台CDH方法,其特征在于:所述步骤S14具体是指:The method for automatically installing a distributed big data platform CDH according to claim 7, wherein the step S14 specifically refers to:判断所述管理节点的配置文件是否按照当前集群机器情况修改IP用户、密码、修改后的机器名;Determine whether the configuration file of the management node has modified the IP user, password, and modified machine name according to the current cluster machine situation;判断所述管理节点目录下是否存在所述配置文件及安装源。Determine whether the configuration file and the installation source exist in the management node directory.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/093339 WO2020258175A1 (en) | 2019-06-27 | 2019-06-27 | Method for automatically installing distributed big data platform cdh |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/093339 WO2020258175A1 (en) | 2019-06-27 | 2019-06-27 | Method for automatically installing distributed big data platform cdh |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020258175A1 true WO2020258175A1 (en) | 2020-12-30 |
Family
ID=74060424
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/093339 WO2020258175A1 (en) | 2019-06-27 | 2019-06-27 | Method for automatically installing distributed big data platform cdh |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2020258175A1 (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102455936A (en) * | 2010-11-25 | 2012-05-16 | 中标软件有限公司 | Rapid Deployment Methods for Clusters |
US20130152078A1 (en) * | 2011-12-09 | 2013-06-13 | Yahoo! Inc. | Method and system for deploying multiple distributed application stacks on a target machine |
CN104317610A (en) * | 2014-10-11 | 2015-01-28 | 福建新大陆软件工程有限公司 | Method and device for automatic installation and deployment of hadoop platform |
CN108549717A (en) * | 2018-04-23 | 2018-09-18 | 泰华智慧产业集团股份有限公司 | The method and system of automatically dispose O&M Hadoop ecology coil assemblies |
-
2019
- 2019-06-27 WO PCT/CN2019/093339 patent/WO2020258175A1/en active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102455936A (en) * | 2010-11-25 | 2012-05-16 | 中标软件有限公司 | Rapid Deployment Methods for Clusters |
US20130152078A1 (en) * | 2011-12-09 | 2013-06-13 | Yahoo! Inc. | Method and system for deploying multiple distributed application stacks on a target machine |
CN104317610A (en) * | 2014-10-11 | 2015-01-28 | 福建新大陆软件工程有限公司 | Method and device for automatic installation and deployment of hadoop platform |
CN108549717A (en) * | 2018-04-23 | 2018-09-18 | 泰华智慧产业集团股份有限公司 | The method and system of automatically dispose O&M Hadoop ecology coil assemblies |
Non-Patent Citations (2)
Title |
---|
SHIRUKAI: "Cloudera Manager Automates the Deployment of CDH Clusters)", 21 November 2018 (2018-11-21), pages 1 - 26, XP009525565, Retrieved from the Internet <URL:https://shirukai.github.io/blog/cloudera-manager-automated-deployment-cdh-cluster.html> * |
XIAOHUI: "CDH5.14 Installation Guide and Maintenance (Set up Many Times in Person)", 6 February 2017 (2017-02-06), pages 1 - 21, XP009525482, Retrieved from the Internet <URL:https://blog.csdn.net/silentwolfyh/article/details/54893826> * |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110286921B (en) | CDH method for automatically installing distributed big data platform | |
US10095496B2 (en) | Single, logical, multi-tier application blueprint used for deployment and management of multiple physical applications in a cloud infrastructure | |
US9710259B2 (en) | System and method for customizing a deployment plan for a multi-tier application in a cloud infrastructure | |
US11941452B2 (en) | System to generate a deployment plan for a cloud infrastructure according to logical, multi-tier application blueprint | |
CN103189839B (en) | The system and method disposed for the automatization of multiunit computer environment | |
US9612817B2 (en) | System and method for providing a physical plugin for use in a cloud platform environment | |
US12032985B2 (en) | Managed virtual machines | |
US10983877B1 (en) | Backup monitoring with automatic verification | |
US10379894B1 (en) | Lineage-based trust for virtual machine images | |
US20230131898A1 (en) | Techniques for building and validating database software in a shared management environment | |
US11966880B2 (en) | Policies and controls for building and validating database software in a shared management environment | |
US20230125904A1 (en) | Recommendation system for building and validating database software in a shared management environment | |
US20220350629A1 (en) | Update management for managed virtual machines | |
WO2023069500A1 (en) | Techniques for building and validating database software in a shared management environment | |
US12118377B2 (en) | Transition to modern management using managed virtual machines | |
US20220350630A1 (en) | Just-in-time assembly for managed virtual machines | |
WO2020258175A1 (en) | Method for automatically installing distributed big data platform cdh | |
Cisco | Initial Installation | |
CN105282180B (en) | Method and device for processing service authentication | |
US20250088417A1 (en) | Architecture for remote delivery and application of datacenter management configurations | |
Kuusisto | Centralized data center provisioning and configuration in managed service provider environment | |
Jaas | Automated Server Infrastructure Deployment | |
Ghonsikar | Openstack cloud installation and uninstallation automation | |
CN116450162A (en) | System and method for deploying and installing interface interactive service components | |
Hwang | Oracle Fusion Middleware Installation Guide for Oracle Data Integrator 11g Release 1 (11.1. 1.7. 0) E16453-06 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19935094 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19935094 Country of ref document: EP Kind code of ref document: A1 |