initial commit

This commit is contained in:
Christoph 2024-04-04 18:19:07 +02:00
commit dde4e8e026
6 changed files with 1992 additions and 0 deletions

View File

@ -0,0 +1,334 @@
# ==========
# Install ElasticSearch
# ==========
# ---
# NOTE:
# we will install a package of ElasticSearch which contains only features
# that are available under the Apache 2.0 license
# ---
# Install 'apt-transport-https' package
#
apt-get install apt-transport-https
# Import the Elasticsearch PGP Key
#
wget -qO - https://artifacts.elastic.co/GPG-KEY-elasticsearch | sudo apt-key add -
# Save the repository definition to '/etc/apt/sources.list.d/elastic-7.x.list'
#
echo "deb https://artifacts.elastic.co/packages/oss-7.x/apt stable main" \
| sudo tee -a /etc/apt/sources.list.d/elastic-7.x.list
# Install the Elasticsearch Debian package with:
apt-get update
apt-get install elasticsearch-oss
# ==========
# Configure ElasticSearch
# ==========
ELASTIC_SEARCH_PUBLISH_IP="0.0.0.0"
ELASTIC_SEARCH_PORT=9200
NODE_NAME="verdi-es"
NODE_NAME="oolm-shop-es"
# ---
# Set System properties
# ---
# Set set sysctl value 'vm.max_map_count' to '524288'
#
# Add to /etc/sysctl.conf:
#
# vm.max_map_count = 524288
#
# Note:
# if installing ElasticSearch into a LX_Container, do this at
# the host system
#
cat << EOF >> /etc/sysctl.conf
# Needed by ElasticSearch Installation on virtual guest
# systems (LX-Containers)
#
# The error message there was:
# max virtual memory areas vm.max_map_count [65530] is too low, increase to at least [262144]
#
vm.max_map_count = 524288
EOF
# ---
# Adjust file '/etc/elasticsearch/elasticsearch.yml'
# ---
# Set network.host to '$ELASTIC_SEARCH_PUBLIC_IP'
#
if ! grep -q -E "^\s*network.host:\s+${ELASTIC_SEARCH_PUBLISH_IP}" /etc/elasticsearch/elasticsearch.yml ; then
if grep -q -E "^\s*#network.host:" /etc/elasticsearch/elasticsearch.yml ; then
perl -i.ORIG -n -p -e "s/^(#network.host:.*)/\1\nnetwork.host: ${ELASTIC_SEARCH_PUBLISH_IP}/" \
/etc/elasticsearch/elasticsearch.yml
else
cat << EOF >> /etc/elasticsearch/elasticsearch.yml
# Additional User Setting (network.host:)
#
network.host: $ELASTIC_SEARCH_PUBLISH_IP
EOF
fi
fi
# Set http.port to '$ELASTIC_SEARCH_PORT'
#
if ! grep -q -E "^\s*http.port:\s+${ELASTIC_SEARCH_PORT}" /etc/elasticsearch/elasticsearch.yml ; then
if grep -q -E "^\s*#http.port:" /etc/elasticsearch/elasticsearch.yml ; then
perl -i.ORIG -n -p -e "s/^(#http.port:.*)/\1\nhttp.port: ${ELASTIC_SEARCH_PORT}/" \
/etc/elasticsearch/elasticsearch.yml
else
cat << EOF >> /etc/elasticsearch/elasticsearch.yml
# Additional User Setting (http.port:)
#
http.port: $ELASTIC_SEARCH_PORT
EOF
fi
fi
# Set values
# node.name: to '$NODE_NAME'
# cluster.initial_master_nodes: to '[${NODE_NAME}]'
#
if ! grep -q -E "^\s*cluster.initial_master_nodes:\s+${NODE_NAME}" \
/etc/elasticsearch/elasticsearch.yml ; then
if grep -q -E "^\s*#cluster.initial_master_nodes:" /etc/elasticsearch/elasticsearch.yml ; then
perl -i.ORIG -n -p -e "s/^(#cluster.initial_master_nodes:.*)/\1\nnode.name: ${NODE_NAME}\ncluster.initial_master_nodes: [\"${NODE_NAME}\"]/" \
/etc/elasticsearch/elasticsearch.yml
else
cat << EOF >> /etc/elasticsearch/elasticsearch.yml
# Additional User Settings:
#
# node.name:
# cluster.initial_master_nodes:
#
node.name: ${NODE_NAME}
cluster.initial_master_nodes: ["${NODE_NAME}"]
EOF
fi
fi
# Activate sysctl settings at file '/etc/sysctl.conf'
#
sysctl -p
# ---
# Note:
# If running as systemd managed process in a LX-Container, the following
# systemd parameters will be set:
# LimitNOFILE=65535
# LimitNPROC=4096
# LimitAS=infinity
# LimitFSIZE=infinity
#
# Take care, your container satisfies this values.
# ---
# ==========
# Test Elasticsearch Setup
# ==========
# The Elasticsearch service is ready to use. You can test it using curl command line
# utility. Run the simple GET command using curl to verify the setup. You will see
# the Elasticsearch cluster details with the version on your screen.
#
# example output:
#
# verdi-es:~ # curl -X GET http://${ELASTIC_SEARCH_PUBLISH_IP}:9200
# {
# "name" : "verdi-es",
# "cluster_name" : "elasticsearch",
# "cluster_uuid" : "J54WIwEqQe203nUbtgOOEA",
# "version" : {
# "number" : "7.2.0",
# "build_flavor" : "oss",
# "build_type" : "deb",
# "build_hash" : "508c38a",
# "build_date" : "2019-06-20T15:54:18.811730Z",
# "build_snapshot" : false,
# "lucene_version" : "8.0.0",
# "minimum_wire_compatibility_version" : "6.8.0",
# "minimum_index_compatibility_version" : "6.0.0-beta1"
# },
# "tagline" : "You Know, for Search"
# }
# verdi-es:~ #
#
curl -X GET http://${ELASTIC_SEARCH_PUBLISH_IP}:9200
# ==========
# Install Kibana (kibana-oss)
# ==========
# Same as Elasticsearch, we will install the latest version of Kibana using the
# apt package manager from the official Elastic repository:
#
apt-get install kibana-oss
# Specifies the address to which the Kibana server will bind. IP addresses and host
# names are both valid values. The default is 'localhost', which usually means
# remote machines will not be able to connect.
# To allow connections from remote users, set this parameter to a non-loopback address.
#
# Set server.host to 'localhost'
#
if ! grep -q -E "^\s*server.host:\s+localhost" /etc/kibana/kibana.yml ; then
if grep -q -E "^\s*#server.host:" /etc/kibana/kibana.yml ; then
perl -i.ORIG -n -p -e "s/^(#server.host:.*)/\1\nserver.host: localhost/" \
/etc/kibana/kibana.yml
else
cat << EOF >> /etc/kibana/kibana.yml
# Additional User Setting (server.host:)
#
server.host: localhost
EOF
fi
fi
# Start the Kibana service and set it to start automatically on boot:
#
systemctl restart kibana
systemctl enable kibana
# ==========
# Configure nginx webserver
# ==========
WEBSITE=verdi-elk.warenform.de
cat <<EOF > /etc/nginx/sites-available/${WEBSITE}.conf
server {
listen 80;
listen [::]:80 ;
server_name ${WEBSITE};
return 301 https://\$host\$request_uri;
}
server {
listen 443 ssl http2;
listen [::]:443 ssl http2;
server_name ${WEBSITE};
root /var/www/html;
index index.html index.htm index.nginx-debian.html;
# Include location directive for Let's Encrypt ACME Challenge
#
# Needed for (automated) updating certificate
#
include snippets/letsencrypt-acme-challenge.conf;
# Diffie-Hellman parameter for DHE ciphersuites, recommended 2048 bits
#
# To generate a dhparam.pem file, run in a terminal
# openssl dhparam -out /etc/nginx/ssl/dhparam.pem 2048
#
ssl_dhparam /etc/nginx/ssl/dhparam.pem;
# Eable session resumption to improve https performance
ssl_session_cache shared:SSL:50m;
ssl_session_timeout 10m;
ssl_session_tickets off;
ssl_ecdh_curve secp384r1;
ssl_certificate /var/lib/dehydrated/certs/${WEBSITE}/fullchain.pem;
ssl_certificate_key /var/lib/dehydrated/certs/${WEBSITE}/privkey.pem;
ssl_protocols TLSv1.2 TLSv1.3;
# ECDHE better than DHE (faster) ECDHE & DHE GCM better than CBC (attacks on AES)
# Everything better than SHA1 (deprecated)
#
#ssl_ciphers 'ECDHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES256-SHA:ECDHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA'
ssl_ciphers ECDHE-RSA-AES256-GCM-SHA512:DHE-RSA-AES256-GCM-SHA512:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-SHA384;
ssl_prefer_server_ciphers on;
#resolver 192.168.42.129 8.8.8.8 valid=300s;
#resolver_timeout 5s;
add_header Strict-Transport-Security "max-age=31536000 always";
add_header X-Frame-Options DENY;
add_header X-Content-Type-Options nosniff;
add_header X-XSS-Protection "1; mode=block";
access_log /var/log/nginx/${WEBSITE}_access.log;
error_log /var/log/nginx/${WEBSITE}_error.log;
auth_basic "Authentication Required";
auth_basic_user_file /etc/nginx/htpasswd.kibana;
location / {
proxy_pass http://localhost:5601;
proxy_http_version 1.1;
proxy_set_header Upgrade \$http_upgrade;
proxy_set_header Connection 'upgrade';
proxy_set_header Host \$host;
proxy_cache_bypass \$http_upgrade;
}
}
EOF
# - Enable site ${WEBSITE}
# -
ln -s ../sites-available/${WEBSITE}.conf /etc/nginx/sites-enabled/
# - Create a basic authentication file with the openssl command:
# -
# - user: admin
# - password: $E%R&T/Z(U
# -
echo "admin:$(openssl passwd -apr1 '$E%R&T/Z(U')" | sudo tee -a /etc/nginx/htpasswd.kiba
# - Restart Nginx Webservice
# -
systemctl restart nginx
# ==========
# Install Logstash (logstash-oss)
# ==========
# The final step is to install Logstash using the apt package manager from
# the official Elastic repository.
#
apt-get install logstash-oss
# Start the Logstash service and set it to start automatically on boot:
#
systemctl restart logstash
systemctl enable logstash
# ---
# Note:
# The Logstash configuration depends on your personal preferences and the
# plugins you will use. You can find more information about how to configure
# Logstash here:
#
# https://www.elastic.co/guide/en/logstash/current/configuration.html
# ---

31
README.JVM-heap-size Normal file
View File

@ -0,0 +1,31 @@
# ==========================
# set JVM heap size manually
# ==========================
# ---
# Set VVM heap size in file /etc/elasticsearch/jvm.options
# ---
...
################################################################
## IMPORTANT: JVM heap size
################################################################
##
## The heap size is automatically configured by Elasticsearch
## based on the available memory in your system and the roles
## each node is configured to fulfill. If specifying heap is
## required, it should be done through a file in jvm.options.d,
## and the min and max should be set to the same value. For
## example, to set the heap to 4 GB, create a new file in the
## jvm.options.d directory containing these lines:
##
## -Xms4g
## -Xmx4g
##
## See https://www.elastic.co/guide/en/elasticsearch/reference/7.17/heap-size.html
## for more information
##
################################################################
-Xms24g
-Xmx24g

327
install_elasticsearch-6.sh Normal file
View File

@ -0,0 +1,327 @@
# ==========
# Install Java
# ==========
# Note:
# Since elasticsearch version 7, java is included in the elastic search package,
# but version 6 DOES NOT contain java.
# Elasticsearch requires at least Java 8 in order to run. It supports both OpenJDK
# and Oracle Java. In this guide, we will install OpenJDK version 8.
#
# We install default-jdk, whi installs (on debian 10) OpenJDK version 11
#
apt-get install default-jdk
# ==========
# Install ElasticSearch
# ==========
# ---
# NOTE:
# we will install a package of ElasticSearch which contains only features
# that are available under the Apache 2.0 license (elasticsearch-oss)
# ---
# Install 'apt-transport-https' package
#
apt-get install apt-transport-https
# Import the Elasticsearch PGP Key
#
wget -qO - https://artifacts.elastic.co/GPG-KEY-elasticsearch | sudo apt-key add -
# Save the repository definition to '/etc/apt/sources.list.d/elastic-7.x.list'
#
echo "deb https://artifacts.elastic.co/packages/oss-6.x/apt stable main" \
| sudo tee -a /etc/apt/sources.list.d/elastic-6.x.list
echo "deb https://artifacts.elastic.co/packages/oss-7.x/apt stable main" \
| sudo tee -a /etc/apt/sources.list.d/elastic-7.x.list
# Install the Elasticsearch Debian package with:
apt-get update
apt-get install elasticsearch-oss
# ==========
# Configure ElasticSearch
# ==========
ELASTIC_SEARCH_PUBLISH_IP="0.0.0.0"
ELASTIC_SEARCH_PORT=9200
# ---
# Set System properties
# ---
# Set set sysctl value 'vm.max_map_count' to '524288'
#
# Add to /etc/sysctl.conf:
#
# vm.max_map_count = 524288
#
# Note:
# if installing ElasticSearch into a LX_Container, do this at
# the host system
#
cat << EOF >> /etc/sysctl.conf
# Needed by ElasticSearch Installation on virtual guest
# systems (LX-Containers)
#
# The error message there was:
# max virtual memory areas vm.max_map_count [65530] is too low, increase to at least [262144]
#
vm.max_map_count = 524288
EOF
# ---
# Adjust file '/etc/elasticsearch/elasticsearch.yml'
# ---
# Set network.host to '$ELASTIC_SEARCH_PUBLIC_IP'
#
if ! grep -q -E "^\s*network.host:\s+${ELASTIC_SEARCH_PUBLISH_IP}" /etc/elasticsearch/elasticsearch.yml ; then
if grep -q -E "^\s*#network.host:" /etc/elasticsearch/elasticsearch.yml ; then
perl -i.ORIG -n -p -e "s/^(#network.host:.*)/\1\nnetwork.host: ${ELASTIC_SEARCH_PUBLISH_IP}/" \
/etc/elasticsearch/elasticsearch.yml
else
cat << EOF >> /etc/elasticsearch/elasticsearch.yml
# Additional User Setting (network.host:)
#
network.host: $ELASTIC_SEARCH_PUBLISH_IP
EOF
fi
fi
# Set http.port to '$ELASTIC_SEARCH_PORT'
#
if ! grep -q -E "^\s*http.port:\s+${ELASTIC_SEARCH_PORT}" /etc/elasticsearch/elasticsearch.yml ; then
if grep -q -E "^\s*#http.port:" /etc/elasticsearch/elasticsearch.yml ; then
perl -i.ORIG -n -p -e "s/^(#http.port:.*)/\1\nhttp.port: ${ELASTIC_SEARCH_PORT}/" \
/etc/elasticsearch/elasticsearch.yml
else
cat << EOF >> /etc/elasticsearch/elasticsearch.yml
# Additional User Setting (http.port:)
#
http.port: $ELASTIC_SEARCH_PORT
EOF
fi
fi
# Activate sysctl settings at file '/etc/sysctl.conf'
#
sysctl -p
# ---
# Note:
# If running as systemd managed process in a LX-Container, the following
# systemd parameters will be set:
# LimitNOFILE=65535
# LimitNPROC=4096
# LimitAS=infinity
# LimitFSIZE=infinity
#
# Take care, your container satisfies this values.
# ---
# ==========
# Test Elasticsearch Setup
# ==========
# The Elasticsearch service is ready to use. You can test it using curl command line
# utility. Run the simple GET command using curl to verify the setup. You will see
# the Elasticsearch cluster details with the version on your screen.
#
# example output:
#
# verdi-es:~ # curl -X GET http://${ELASTIC_SEARCH_PUBLISH_IP}:9200
# {
# "name" : "verdi-es",
# "cluster_name" : "elasticsearch",
# "cluster_uuid" : "J54WIwEqQe203nUbtgOOEA",
# "version" : {
# "number" : "7.2.0",
# "build_flavor" : "oss",
# "build_type" : "deb",
# "build_hash" : "508c38a",
# "build_date" : "2019-06-20T15:54:18.811730Z",
# "build_snapshot" : false,
# "lucene_version" : "8.0.0",
# "minimum_wire_compatibility_version" : "6.8.0",
# "minimum_index_compatibility_version" : "6.0.0-beta1"
# },
# "tagline" : "You Know, for Search"
# }
# verdi-es:~ #
#
curl -X GET http://${ELASTIC_SEARCH_PUBLISH_IP}:9200
# ==========
# Install Kibana (kibana-oss)
# ==========
# Same as Elasticsearch, we will install the latest version of Kibana using the
# apt package manager from the official Elastic repository:
#
apt-get install kibana-oss
# Specifies the address to which the Kibana server will bind. IP addresses and host
# names are both valid values. The default is 'localhost', which usually means
# remote machines will not be able to connect.
# To allow connections from remote users, set this parameter to a non-loopback address.
#
# Set server.host to 'localhost'
#
if ! grep -q -E "^\s*server.host:\s+localhost" /etc/kibana/kibana.yml ; then
if grep -q -E "^\s*#server.host:" /etc/kibana/kibana.yml ; then
perl -i.ORIG -n -p -e "s/^(#server.host:.*)/\1\nserver.host: localhost/" \
/etc/kibana/kibana.yml
else
cat << EOF >> /etc/kibana/kibana.yml
# Additional User Setting (server.host:)
#
server.host: localhost
EOF
fi
fi
# Start the Kibana service and set it to start automatically on boot:
#
systemctl restart kibana
systemctl enable kibana
# ==========
# Configure nginx webserver
# ==========
WEBSITE=verdi-elk.warenform.de
cat <<EOF > /etc/nginx/sites-available/${WEBSITE}.conf
server {
listen 80;
listen [::]:80 ;
server_name ${WEBSITE};
return 301 https://\$host\$request_uri;
}
server {
listen 443 ssl http2;
listen [::]:443 ssl http2;
server_name ${WEBSITE};
root /var/www/html;
index index.html index.htm index.nginx-debian.html;
# Include location directive for Let's Encrypt ACME Challenge
#
# Needed for (automated) updating certificate
#
include snippets/letsencrypt-acme-challenge.conf;
# Diffie-Hellman parameter for DHE ciphersuites, recommended 2048 bits
#
# To generate a dhparam.pem file, run in a terminal
# openssl dhparam -out /etc/nginx/ssl/dhparam.pem 2048
#
ssl_dhparam /etc/nginx/ssl/dhparam.pem;
# Eable session resumption to improve https performance
ssl_session_cache shared:SSL:50m;
ssl_session_timeout 10m;
ssl_session_tickets off;
ssl_ecdh_curve secp384r1;
ssl_certificate /var/lib/dehydrated/certs/${WEBSITE}/fullchain.pem;
ssl_certificate_key /var/lib/dehydrated/certs/${WEBSITE}/privkey.pem;
ssl_protocols TLSv1.2 TLSv1.3;
# ECDHE better than DHE (faster) ECDHE & DHE GCM better than CBC (attacks on AES)
# Everything better than SHA1 (deprecated)
#
#ssl_ciphers 'ECDHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES256-SHA:ECDHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA'
ssl_ciphers ECDHE-RSA-AES256-GCM-SHA512:DHE-RSA-AES256-GCM-SHA512:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-SHA384;
ssl_prefer_server_ciphers on;
#resolver 192.168.42.129 8.8.8.8 valid=300s;
#resolver_timeout 5s;
add_header Strict-Transport-Security "max-age=31536000 always";
add_header X-Frame-Options DENY;
add_header X-Content-Type-Options nosniff;
add_header X-XSS-Protection "1; mode=block";
access_log /var/log/nginx/${WEBSITE}_access.log;
error_log /var/log/nginx/${WEBSITE}_error.log;
auth_basic "Authentication Required";
auth_basic_user_file /etc/nginx/htpasswd.kibana;
location / {
proxy_pass http://localhost:5601;
proxy_http_version 1.1;
proxy_set_header Upgrade \$http_upgrade;
proxy_set_header Connection 'upgrade';
proxy_set_header Host \$host;
proxy_cache_bypass \$http_upgrade;
}
}
EOF
# - Enable site ${WEBSITE}
# -
ln -s ../sites-available/${WEBSITE}.conf /etc/nginx/sites-enabled/
# - Create a basic authentication file with the openssl command:
# -
# - user: admin
# - password: $E%R&T/Z(U
# -
echo "admin:$(openssl passwd -apr1 '$E%R&T/Z(U')" | sudo tee -a /etc/nginx/htpasswd.kiba
# - Restart Nginx Webservice
# -
systemctl restart nginx
# ==========
# Install Logstash (logstash-oss)
# ==========
# The final step is to install Logstash using the apt package manager from
# the official Elastic repository.
#
apt-get install logstash-oss
# Start the Logstash service and set it to start automatically on boot:
#
systemctl restart logstash
systemctl enable logstash
# ---
# Note:
# The Logstash configuration depends on your personal preferences and the
# plugins you will use. You can find more information about how to configure
# Logstash here:
#
# https://www.elastic.co/guide/en/logstash/current/configuration.html
# ---

View File

@ -0,0 +1,2 @@
# see:
# https://wiki.crowncloud.net/How_to_Install_Elasticsearch_on_Debian_12?How_to_Install_Elasticsearch_on_Debian_12

667
install_elasticsearch-7.sh Normal file
View File

@ -0,0 +1,667 @@
# ==========
# Install ElasticSearch
# ==========
# ---
# NOTE:
# we will install a package of ElasticSearch which contains only features
# that are available under the Apache 2.0 license
# ---
# Install 'apt-transport-https' package
#
apt-get install apt-transport-https
# Import the Elasticsearch PGP Key
#
# We sign all of our packages with the Elasticsearch Signing Key (PGP key D88E42B4,
# available from https://pgp.mit.edu) with fingerprint:
#
# 4609 5ACC 8548 582C 1A26 99A9 D27D 666C D88E 42B4
#
wget -qO - https://artifacts.elastic.co/GPG-KEY-elasticsearch | sudo gpg --dearmor -o /usr/share/keyrings/elasticsearch-keyring.gpg
# You may need to install the apt-transport-https package on Debian before proceeding:
#
apt-get install apt-transport-https
# Save the repository definition to '/etc/apt/sources.list.d/elastic-7.x.list'
#
echo "deb [signed-by=/usr/share/keyrings/elasticsearch-keyring.gpg] https://artifacts.elastic.co/packages/7.x/apt stable main" | sudo tee /etc/apt/sources.list.d/elastic-7.x.list
# Install the Elasticsearch Debian package with:
apt-get update
apt-get install elasticsearch
# ==========
# Configure ElasticSearch
# ==========
# By default Elasticsearch is only accessible on localhost.
#
# Leave empty for accepting the default
#
ELASTIC_SEARCH_PUBLISH_IP="0.0.0.0"
ELASTIC_SEARCH_PUBLISH_IP=""
ELASTIC_SEARCH_PUBLISH_IP="127.0.0.1"
# By default Elasticsearch listens for HTTP traffic on the first free port it
# finds starting at 9200. Set a specific HTTP port here:
#
# Leave empty for accepting the default
#
ELASTIC_SEARCH_PORT=9200
NODE_NAME="verdi-es"
NODE_NAME="oolm-shop-es"
DISCOVERY_TYPE='single-node'
AUTOMATIC_CREATION_SYSTEM_INDICES=false
ELASTIC_SEARCH_CONFIG_FILE="/etc/elasticsearch/elasticsearch.yml"
ELASTIC_SEARCH_CONFIG_FILE="/root/elasticsearch.yml"
# ---
# Set System properties
# ---
# Set set sysctl value 'vm.max_map_count' to '524288'
#
# Add to /etc/sysctl.conf:
#
# vm.max_map_count = 524288
#
# Note:
# if installing ElasticSearch into a LX_Container, do this at
# the host system
#
cat << EOF >> /etc/sysctl.d/60-elasticsearch.conf
# Needed by ElasticSearch Installation on virtual guest
# systems (LX-Containers)
#
# The error message there was:
# max virtual memory areas vm.max_map_count [65530] is too low, increase to at least [262144]
#
vm.max_map_count = 524288
EOF
# Activate sysctl settings at file '/etc/sysctl.conf'
#
sysctl -p
# ---
# Adjust file '/etc/elasticsearch/elasticsearch.yml' - ${ELASTIC_SEARCH_CONFIG_FILE}
# ---
if [[ ! -f "${ELASTIC_SEARCH_CONFIG_FILE}.ORIG" ]] ; then
cp -a "${ELASTIC_SEARCH_CONFIG_FILE}" "${ELASTIC_SEARCH_CONFIG_FILE}.ORIG"
fi
# Set network.host to '$ELASTIC_SEARCH_PUBLIC_IP'
#
# By default Elasticsearch is only accessible on localhost. Set a different
# address here to expose this node on the network:
#
# Replce only first occurence of match
#
if [[ -n "$ELASTIC_SEARCH_PUBLISH_IP" ]]; then
if ! grep -q -E "^\s*network.host:\s+${ELASTIC_SEARCH_PUBLISH_IP}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#network.host:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#network.host:.*\)/ s//\1\nnetwork.host: ${ELASTIC_SEARCH_PUBLISH_IP}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# network.host
#
# address here to expose this node on the network:
#
network.host: $ELASTIC_SEARCH_PUBLISH_IP
EOF
fi
fi
fi
# Set http.port to '$ELASTIC_SEARCH_PORT'
#
# Replce only first occurence of match
#
if [[ -n "$ELASTIC_SEARCH_PORT" ]]; then
if ! grep -q -E "^\s*http.port:\s+${ELASTIC_SEARCH_PORT}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#http.port:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#http.port:.*\)/ s//\1\nhttp.port: ${ELASTIC_SEARCH_PORT}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# http.port
#
# By default Elasticsearch listens for HTTP traffic on the first free port it
# finds starting at 9200. Set a specific HTTP port here:
#
http.port: $ELASTIC_SEARCH_PORT
EOF
fi
fi
fi
# Set values
# node.name: to '$NODE_NAME'
#
# # if discovery.type IS NOT set to 'single-node'
# cluster.initial_master_nodes: to '[${NODE_NAME}]'
#
if [[ -n "$NODE_NAME" ]]; then
if ! grep -q -E "^\s*node.name:\s+${NODE_NAME}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#node.name:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#node.name.*\)/ s//\1\nnode.name: ${NODE_NAME}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# node.name
#
# Use a descriptive name for the node:
#
node.name: ${NODE_NAME}
EOF
if [[ -n "${DISCOVERY_TYPE}" ]] && [[ "single-node" != "${DISCOVERY_TYPE}" ]] ; then
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
cluster.initial_master_nodes: ["${NODE_NAME}"]
EOF
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# setting [cluster.initial_master_nodes] is not allowed when [discovery.type] is set to [single-node]
#cluster.initial_master_nodes: ["${NODE_NAME}"]
EOF
fi
fi
fi
if ! grep -q -E "^\s*cluster.initial_master_nodes:\s+${NODE_NAME}" \
${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#cluster.initial_master_nodes:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if [[ -n "${DISCOVERY_TYPE}" ]] && [[ "single-node" != "${DISCOVERY_TYPE}" ]] ; then
sed -i "0,/^\(#cluster.initial_master_nodes.*\)/ s//\1\ncluster.initial_master_nodes: [\"${NODE_NAME}\"]/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
sed -i "0,/^\(#cluster.initial_master_nodes.*\)/ s//\1\n#cluster.initial_master_nodes: [\"${NODE_NAME}\"]/" \
${ELASTIC_SEARCH_CONFIG_FILE}
fi
elif grep -q -E "^\s*cluster.initial_master_nodes:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if [[ -n "${DISCOVERY_TYPE}" ]] && [[ "single-node" != "${DISCOVERY_TYPE}" ]] ; then
sed -i "0,/^\(cluster.initial_master_nodes.*\)/ s//#\1\ncluster.initial_master_nodes: [\"${NODE_NAME}\"]/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
sed -i "0,/^\(cluster.initial_master_nodes.*\)/ s//#\1/" ${ELASTIC_SEARCH_CONFIG_FILE}
fi
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# cluster.initial_master_nodes:
#
EOF
if [[ -n "${DISCOVERY_TYPE}" ]] && [[ "single-node" != "${DISCOVERY_TYPE}" ]] ; then
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
cluster.initial_master_nodes: ["${NODE_NAME}"]
EOF
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# setting [cluster.initial_master_nodes] is not allowed when [discovery.type] is set to [single-node]
#cluster.initial_master_nodes: ["${NODE_NAME}"]
EOF
fi
fi
fi
fi
# Set value discovery.type
#
if [[ -n "$DISCOVERY_TYPE" ]]; then
if ! grep -q -E "^\s*discovery.type:\s+${DISCOVERY_TYPE}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#discovery.type:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#discovery.type.*\)/ s//\1\ndiscovery.type: ${DISCOVERY_TYPE}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
elif grep -q -E "^\s*discovery.type:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(discovery.type.*\)/ s//#\1\ndiscovery.type: ${DISCOVERY_TYPE}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# discovery.type
#
# (Static) Specifies whether Elasticsearch should form a multiple-node cluster.
#
# Defaults to multi-node, which means that Elasticsearch discovers other nodes
# when forming a cluster and allows other nodes to join the cluster later.
#
# If set to single-node, Elasticsearch forms a single-node cluster and suppresses
# the timeout set by cluster.publish.timeout. For more information about when
# you might use this setting, see Single-node discovery:
#
# https://www.elastic.co/guide/en/elasticsearch/reference/current/important-settings.html#initial_master_nodes
#
#
# (Statisch) Gibt an, ob Elasticsearch einen Mehrknoten-Cluster bilden soll.
#
# Der Standardwert ist Multi-Node, was bedeutet, dass Elasticsearch andere Knoten
# bei der Bildung eines Clusters entdeckt und anderen Knoten erlaubt, dem Cluster
# später beizutreten.
#
# Wenn auf single-node gesetzt, bildet Elasticsearch einen Single-Node-Cluster und
# unterdrückt das Timeout, das durch cluster.publish.timeout gesetzt wird. Für
# weitere Informationen darüber, wann diese Einstellung verwendet werden kann,
# siehe Single-node discovery:
#
# https://www.elastic.co/guide/en/elasticsearch/reference/current/important-settings.html#initial_master_nodes
#
discovery.type: ${DISCOVERY_TYPE}
EOF
fi
fi
fi
# The default behavior of disabling security on basic licenses is deprecated.
# In a later version of Elasticsearch, the value of [xpack.security.enabled]
# will default to "true" , regardless of the license level.
# See
# https://www.elastic.co/guide/en/elasticsearch/reference/7.17/security-minimal-setup.html
# to enable security, or explicitly disable security by setting
# [xpack.security.enabled] to false in elasticsearch.yml
#
# See
# https://www.elastic.co/guide/en/elasticsearch/reference/7.17/security-minimal-setup.html
# to enable security.
#
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# See
# https://www.elastic.co/guide/en/elasticsearch/reference/7.17/security-minimal-setup.html
# to enable security.
#
xpack.security.enabled: false
EOF
# Disable geoip usage
#
# Maybe or better hopefully avoid error
#
# exception during geoip databases update..
#
# see also: https://discuss.elastic.co/t/how-to-disable-geoip-usage-in-7-14-0/281076/8
#
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# Disable geoip usage
#
# exception during geoip databases update..
#
# see also: https://discuss.elastic.co/t/how-to-disable-geoip-usage-in-7-14-0/281076/8
#
ingest.geoip.downloader.enabled: false
EOF
# Enable automatic creation of system indices
#
# Some commercial features automatically create indices within Elasticsearch. By default,
# Elasticsearch is configured to allow automatic index creation, and no additional steps
# are required. However, if you have disabled automatic index creation in Elasticsearch,
# you must configure action.auto_create_index in elasticsearch.yml to allow the commercial
# features to create the following indices:
#
if $AUTOMATIC_CREATION_SYSTEM_INDICES ; then
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# Additional User Settings:
#
# action.auto_create_index
#
# Enable automatic creation of system indices
#
# Some commercial features automatically create indices within Elasticsearch. By default,
# Elasticsearch is configured to allow automatic index creation, and no additional steps
# are required. However, if you have disabled automatic index creation in Elasticsearch,
# you must configure action.auto_create_index in elasticsearch.yml to allow the commercial
# features to create the following indices:
#
action.auto_create_index: .monitoring*,.watches,.triggered_watches,.watcher-history*,.ml*
EOF
fi
# ---
# Note:
# If running as systemd managed process in a LX-Container, the following
# systemd parameters will be set:
# LimitNOFILE=65535
# LimitNPROC=4096
# LimitAS=infinity
# LimitFSIZE=infinity
#
# Take care, your container satisfies this values.
# ---
# ==========
# Running Elasticsearch with systemd
# ==========
# To configure Elasticsearch to start automatically when the system boots up, run
# the following commands:
#
systemctl daemon-reload
systemctl enable elasticsearch.service
# Elasticsearch can be started and stopped as follows:
#
systemctl start elasticsearch.service
systemctl stop elasticsearch.service
# If you have password-protected your Elasticsearch keystore, you will need to
# provide systemd with the keystore password using a local file and systemd
# environment variables. This local file should be protected while it exists and
# may be safely deleted once Elasticsearch is up and running.
#
KEYSTORE_PASSPHRASE_FILE=</path/to/my_pwd_file.tmp>
systemctl stop elasticsearch.service
echo "keystore_password" > $KEYSTORE_PASSPHRASE_FILE
chmod 600 $KEYSTORE_PASSPHRASE_FILE
systemctl set-environment ES_KEYSTORE_PASSPHRASE_FILE=$KEYSTORE_PASSPHRASE_FILE
systemctl start elasticsearch.service
# ==========
# Test Elasticsearch Setup
# ==========
# The Elasticsearch service is ready to use. You can test it using curl command line
# utility. Run the simple GET command using curl to verify the setup. You will see
# the Elasticsearch cluster details with the version on your screen.
#
# example output:
#
# verdi-es:~ # curl -X GET http://${ELASTIC_SEARCH_PUBLISH_IP}:9200
# {
# "name" : "verdi-es",
# "cluster_name" : "elasticsearch",
# "cluster_uuid" : "J54WIwEqQe203nUbtgOOEA",
# "version" : {
# "number" : "7.2.0",
# "build_flavor" : "oss",
# "build_type" : "deb",
# "build_hash" : "508c38a",
# "build_date" : "2019-06-20T15:54:18.811730Z",
# "build_snapshot" : false,
# "lucene_version" : "8.0.0",
# "minimum_wire_compatibility_version" : "6.8.0",
# "minimum_index_compatibility_version" : "6.0.0-beta1"
# },
# "tagline" : "You Know, for Search"
# }
# verdi-es:~ #
#
curl -X GET http://${ELASTIC_SEARCH_PUBLISH_IP}:9200
# ==========
# Install Kibana
# ==========
KIBANA_SERVER_HOST="localhost"
KIBANA_SERVER_PORT="5601"
# Same as Elasticsearch, we will install the latest version of Kibana using the
# apt package manager from the official Elastic repository:
#
apt-get install kibana
# ---
# Adjust file '/etc/kibana/kibana.yml'
# ---
# see also: https://www.elastic.co/guide/en/kibana/8.3/settings.html
if [[ ! -f "/etc/kibana/kibana.yml.ORIG" ]] ; then
cp -a "/etc/kibana/kibana.yml" "/etc/kibana/kibana.yml.ORIG"
fi
# Specifies the address to which the Kibana server will bind. IP addresses and host
# names are both valid values. The default is 'localhost', which usually means
# remote machines will not be able to connect.
# To allow connections from remote users, set this parameter to a non-loopback address.
#
# Set server.host to 'localhost'
#
if ! grep -q -E "^\s*server.host:\s+${KIBANA_SERVER_HOST}" /etc/kibana/kibana.yml ; then
if grep -q -E "^\s*#server.host:" /etc/kibana/kibana.yml ; then
perl -i -n -p -e "s/^(#server.host:.*)/\1\nserver.host: ${KIBANA_SERVER_HOST}/" \
/etc/kibana/kibana.yml
else
cat << EOF >> /etc/kibana/kibana.yml
# Additional User Setting (server.host:)
#
server.host: ${KIBANA_SERVER_HOST}
EOF
fi
fi
# Set server.port to '5601'
#
if ! grep -q -E "^\s*server.port:\s+${KIBANA_SERVER_PORT}" /etc/kibana/kibana.yml ; then
if grep -q -E "^\s*#server.port:" /etc/kibana/kibana.yml ; then
perl -i -n -p -e "s/^(#server.port:.*)/\1\nserver.port: ${KIBANA_SERVER_PORT}/" \
/etc/kibana/kibana.yml
else
cat << EOF >> /etc/kibana/kibana.yml
# Additional User Setting (server.port:)
#
server.port: ${KIBANA_SERVER_HOST}
EOF
fi
fi
# ---
# Run Kibana with systemd
# ---
# Configure Kibana to start automatically when the system starts:
#
systemctl daemon-reload
systemctl enable kibana.service
# Kibana service ican be started and stopped as follows:
##
systemctl start kibana.service
systemctl istop kibana.service
# ==========
# Configure nginx webserver
# ==========
WEBSITE=verdi-elk.warenform.de
cat <<EOF > /etc/nginx/sites-available/${WEBSITE}.conf
server {
listen 80;
listen [::]:80 ;
server_name ${WEBSITE};
return 301 https://\$host\$request_uri;
}
server {
listen 443 ssl http2;
listen [::]:443 ssl http2;
server_name ${WEBSITE};
root /var/www/html;
index index.html index.htm index.nginx-debian.html;
# Include location directive for Let's Encrypt ACME Challenge
#
# Needed for (automated) updating certificate
#
include snippets/letsencrypt-acme-challenge.conf;
# Diffie-Hellman parameter for DHE ciphersuites, recommended 2048 bits
#
# To generate a dhparam.pem file, run in a terminal
# openssl dhparam -out /etc/nginx/ssl/dhparam.pem 2048
#
ssl_dhparam /etc/nginx/ssl/dhparam.pem;
# Eable session resumption to improve https performance
ssl_session_cache shared:SSL:50m;
ssl_session_timeout 10m;
ssl_session_tickets off;
ssl_ecdh_curve secp384r1;
ssl_certificate /var/lib/dehydrated/certs/${WEBSITE}/fullchain.pem;
ssl_certificate_key /var/lib/dehydrated/certs/${WEBSITE}/privkey.pem;
ssl_protocols TLSv1.2 TLSv1.3;
# ECDHE better than DHE (faster) ECDHE & DHE GCM better than CBC (attacks on AES)
# Everything better than SHA1 (deprecated)
#
#ssl_ciphers 'ECDHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES256-SHA:ECDHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA'
ssl_ciphers ECDHE-RSA-AES256-GCM-SHA512:DHE-RSA-AES256-GCM-SHA512:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-SHA384;
ssl_prefer_server_ciphers on;
#resolver 192.168.42.129 8.8.8.8 valid=300s;
#resolver_timeout 5s;
add_header Strict-Transport-Security "max-age=31536000 always";
add_header X-Frame-Options DENY;
add_header X-Content-Type-Options nosniff;
add_header X-XSS-Protection "1; mode=block";
access_log /var/log/nginx/${WEBSITE}_access.log;
error_log /var/log/nginx/${WEBSITE}_error.log;
auth_basic "Authentication Required";
auth_basic_user_file /etc/nginx/htpasswd.kibana;
location / {
proxy_pass http://localhost:5601;
proxy_http_version 1.1;
proxy_set_header Upgrade \$http_upgrade;
proxy_set_header Connection 'upgrade';
proxy_set_header Host \$host;
proxy_cache_bypass \$http_upgrade;
}
}
EOF
# - Enable site ${WEBSITE}
# -
ln -s ../sites-available/${WEBSITE}.conf /etc/nginx/sites-enabled/
# - Create a basic authentication file with the openssl command:
# -
# - user: admin
# - password: $E%R&T/Z(U
# -
echo "admin:$(openssl passwd -apr1 '$E%R&T/Z(U')" | sudo tee -a /etc/nginx/htpasswd.kiba
# - Restart Nginx Webservice
# -
systemctl restart nginx
# ==========
# Install Logstash (logstash)
# ==========
# The final step is to install Logstash using the apt package manager from
# the official Elastic repository.
#
apt-get install logstash
# Start the Logstash service and set it to start automatically on boot:
#
systemctl restart logstash
systemctl enable logstash
# ---
# Note:
# The Logstash configuration depends on your personal preferences and the
# plugins you will use. You can find more information about how to configure
# Logstash here:
#
# https://www.elastic.co/guide/en/logstash/current/configuration.html
# ---

631
install_elasticsearch-8.sh Normal file
View File

@ -0,0 +1,631 @@
# ==========
# Install ElasticSearch
# ==========
# see: https://www.elastic.co/guide/en/elasticsearch/reference/current/deb.html
# ---
# NOTE:
# we will install a package of ElasticSearch which contains only features
# that are available under the Apache 2.0 license
# ---
# Install 'apt-transport-https' package
#
apt-get install apt-transport-https
# Import the Elasticsearch PGP Key
#
wget -qO - https://artifacts.elastic.co/GPG-KEY-elasticsearch | sudo gpg --dearmor -o /usr/share/keyrings/elasticsearch-keyring.gpg
# Save the repository definition to '/etc/apt/sources.list.d/elastic-7.x.list'
#
echo "deb [signed-by=/usr/share/keyrings/elasticsearch-keyring.gpg] https://artifacts.elastic.co/packages/8.x/apt stable main" | sudo tee /etc/apt/sources.list.d/elastic-8.x.list
# Install the Elasticsearch Debian package with:
#
# !! NOTE !!
#
# When installing Elasticsearch, security features are enabled and configured by default.
# When you install Elasticsearch, the following security configuration occurs automatically:
#
# - Authentication and authorization are enabled, and a password is generated for the
# elastic built-in superuser.
#
# - Certificates and keys for TLS are generated for the transport and HTTP layer, and
# TLS is enabled and configured with these keys and certificates.
#
# The password and certificate and keys are output to your terminal.
# ==================================================================
#
apt-get update
if [[ -f "/root/apt-get_install_elasticsearch.out" ]] ; then
cp -a "/root/apt-get_install_elasticsearch.out" "/root/apt-get_install_elasticsearch.out.$(date +%Y-%m-%d-%H%M)"
fi
apt-get install elasticsearch | tee /root/apt-get_install_elasticsearch.out
# ==========
# Configure ElasticSearch
# ==========
# By default Elasticsearch is only accessible on localhost.
#
# Leave empty for accepting the default
#
ELASTIC_SEARCH_PUBLISH_IP="0.0.0.0"
ELASTIC_SEARCH_PUBLISH_IP=""
# By default Elasticsearch listens for HTTP traffic on the first free port it
# finds starting at 9200. Set a specific HTTP port here:
#
# Leave empty for accepting the default
#
ELASTIC_SEARCH_PORT=9200
NODE_NAME="verdi-es"
NODE_NAME="oolm-shop-es"
DISCOVERY_TYPE='single-node'
ELASTIC_SEARCH_CONFIG_FILE="/etc/elasticsearch/elasticsearch.yml"
ELASTIC_SEARCH_CONFIG_FILE="/root/elasticsearch.yml"
# ---
# Set System properties
# ---
# Set set sysctl value 'vm.max_map_count' to '524288'
#
# Add to /etc/sysctl.conf:
#
# vm.max_map_count = 524288
#
# Note:
# if installing ElasticSearch into a LX_Container, do this at
# the host system
#
cat << EOF >> /etc/sysctl.d/60-elasticsearch.conf
# Needed by ElasticSearch Installation on virtual guest
# systems (LX-Containers)
#
# The error message there was:
# max virtual memory areas vm.max_map_count [65530] is too low, increase to at least [262144]
#
vm.max_map_count = 524288
EOF
# Activate sysctl settings at file '/etc/sysctl.conf'
#
sysctl -p
# ---
# Adjust file '${ELASTIC_SEARCH_CONFIG_FILE}'
# ---
if [[ ! -f "${ELASTIC_SEARCH_CONFIG_FILE}.ORIG" ]] ; then
cp -a "${ELASTIC_SEARCH_CONFIG_FILE}" "${ELASTIC_SEARCH_CONFIG_FILE}.ORIG"
fi
# Set network.host to '$ELASTIC_SEARCH_PUBLIC_IP'
#
# By default Elasticsearch is only accessible on localhost. Set a different
# address here to expose this node on the network:
#
# Replce only first occurence of match
#
if [[ -n "$ELASTIC_SEARCH_PUBLISH_IP" ]]; then
if ! grep -q -E "^\s*network.host:\s+${ELASTIC_SEARCH_PUBLISH_IP}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#network.host:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#network.host:.*\)/ s//\1\nnetwork.host: ${ELASTIC_SEARCH_PUBLISH_IP}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# network.host
#
# address here to expose this node on the network:
#
EOF
fi
fi
fi
# Set http.port to '$ELASTIC_SEARCH_PORT'
#
# Replce only first occurence of match
#
if [[ -n "$ELASTIC_SEARCH_PORT" ]]; then
if ! grep -q -E "^\s*http.port:\s+${ELASTIC_SEARCH_PORT}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#http.port:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#http.port:.*\)/ s//\1\nhttp.port: ${ELASTIC_SEARCH_PORT}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# http.port
#
# By default Elasticsearch listens for HTTP traffic on the first free port it
# finds starting at 9200. Set a specific HTTP port here:
#
http.port: $ELASTIC_SEARCH_PORT
EOF
fi
fi
fi
# Set values
# node.name: to '$NODE_NAME'
# cluster.initial_master_nodes: to '[${NODE_NAME}]'
#
if [[ -n "$NODE_NAME" ]]; then
if ! grep -q -E "^\s*node.name:\s+${NODE_NAME}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#node.name:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#node.name.*\)/ s//\1\nnode.name: ${NODE_NAME}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# node.name
#
# Use a descriptive name for the node:
#
node.name: ${NODE_NAME}
cluster.initial_master_nodes: ["${NODE_NAME}"]
EOF
fi
fi
if ! grep -q -E "^\s*cluster.initial_master_nodes:\s+${NODE_NAME}" \
${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#cluster.initial_master_nodes:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#cluster.initial_master_nodes.*\)/ s//\1\ncluster.initial_master_nodes: [\"${NODE_NAME}\"]/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# cluster.initial_master_nodes:
#
cluster.initial_master_nodes: ["${NODE_NAME}"]
EOF
fi
fi
fi
if [[ -n "$DISCOVERY_TYPE" ]]; then
if ! grep -q -E "^\s*discovery.type:\s+${DISCOVERY_TYPE}" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
if grep -q -E "^\s*#discovery.type:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(#discovery.type.*\)/ s//\1\ndiscovery.type: ${DISCOVERY_TYPE}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
elif grep -q -E "^\s*discovery.type:" ${ELASTIC_SEARCH_CONFIG_FILE} ; then
sed -i "0,/^\(discovery.type.*\)/ s//#\1\ndiscovery.type: ${DISCOVERY_TYPE}/" \
${ELASTIC_SEARCH_CONFIG_FILE}
else
cat << EOF >> ${ELASTIC_SEARCH_CONFIG_FILE}
# ----------
# Additional User Setting:
# ----------
# discovery.type
#
# (Static) Specifies whether Elasticsearch should form a multiple-node cluster.
#
# Defaults to multi-node, which means that Elasticsearch discovers other nodes
# when forming a cluster and allows other nodes to join the cluster later.
#
# If set to single-node, Elasticsearch forms a single-node cluster and suppresses
# the timeout set by cluster.publish.timeout. For more information about when
# you might use this setting, see Single-node discovery:
#
# https://www.elastic.co/guide/en/elasticsearch/reference/current/important-settings.html#initial_master_nodes
#
#
# (Statisch) Gibt an, ob Elasticsearch einen Mehrknoten-Cluster bilden soll.
#
# Der Standardwert ist Multi-Node, was bedeutet, dass Elasticsearch andere Knoten
# bei der Bildung eines Clusters entdeckt und anderen Knoten erlaubt, dem Cluster
# später beizutreten.
#
# Wenn auf single-node gesetzt, bildet Elasticsearch einen Single-Node-Cluster und
# unterdrückt das Timeout, das durch cluster.publish.timeout gesetzt wird. Für
# weitere Informationen darüber, wann diese Einstellung verwendet werden kann,
# siehe Single-node discovery:
#
# https://www.elastic.co/guide/en/elasticsearch/reference/current/important-settings.html#initial_master_nodes
#
discovery.type: ${DISCOVERY_TYPE}
EOF
fi
fi
fi
# ---
# Note:
# If running as systemd managed process in a LX-Container, the following
# systemd parameters will be set:
# LimitNOFILE=65535
# LimitNPROC=4096
# LimitAS=infinity
# LimitFSIZE=infinity
#
# Take care, your container satisfies this values.
# ---
# ==========
# Reconfigure a node to join an existing cluster
# ==========
# When you install Elasticsearch, the installation process configures
# a single-node cluster by default. If you want a node to join an existing
# cluster instead, generate an enrollment token on an existing node before
# you start the new node for the first time.
#
# 1. On any node in your existing cluster, generate a node enrollment token:
#
/usr/share/elasticsearch/bin/elasticsearch-create-enrollment-token -s node
# 2. Copy the enrollment token, which is output to your terminal.
#
# 3. On your new Elasticsearch node, pass the enrollment token as a parameter
# to the elasticsearch-reconfigure-node tool:
#
/usr/share/elasticsearch/bin/elasticsearch-reconfigure-node --enrollment-token <enrollment-token>
# 4. Start your new node using systemd.
# ==========
# Enable automatic creation of system indices
# ==========
# Some commercial features automatically create indices within Elasticsearch.
# By default, Elasticsearch is configured to allow automatic index creation,
# and no additional steps are required. However, if you have disabled automatic
# index creation in Elasticsearch, you must configure action.auto_create_index
# in elasticsearch.yml to allow the commercial features to create the following
# indices:
#
action.auto_create_index: .monitoring*,.watches,.triggered_watches,.watcher-history*,.ml*
# ==========
# Running Elasticsearch with systemd
# ==========
# To configure Elasticsearch to start automatically when the system boots up, run
# the following commands:
#
systemctl daemon-reload
systemctl enable elasticsearch.service
# Elasticsearch can be started and stopped as follows:
#
systemctl start elasticsearch.service
systemctl stop elasticsearch.service
# !! NOTE !!
#
# If you have password-protected your Elasticsearch keystore, you will need to provide
# systemd with the keystore password using a local file and systemd environment
# variables. This local file should be protected while it exists and may be safely
# deleted once Elasticsearch is up and running.
#
KEYSTORE_PASSWORD="<keystore_password>"
KEYSTORE_PASSWORD_FILE="</path/to/my_pwd_file.tmp>"
echo "${KEYSTORE_PASSWORD}" > "${KEYSTORE_PASSWORD_FILE}"
chmod 600 "${KEYSTORE_PASSWORD_FILE}"
systemctl set-environment ES_KEYSTORE_PASSPHRASE_FILE=${KEYSTORE_PASSWORD_FILE}
systemctl start elasticsearch.service
# ==========
# Test Elasticsearch Setup
# ==========
# The Elasticsearch service is ready to use. You can test it using curl command line
# utility. Run the simple GET command using curl to verify the setup. You will see
# the Elasticsearch cluster details with the version on your screen.
#
# example output:
#
# verdi-es:~ # curl -X GET http://${ELASTIC_SEARCH_PUBLISH_IP}:9200
# {
# "name" : "verdi-es",
# "cluster_name" : "elasticsearch",
# "cluster_uuid" : "J54WIwEqQe203nUbtgOOEA",
# "version" : {
# "number" : "7.2.0",
# "build_flavor" : "oss",
# "build_type" : "deb",
# "build_hash" : "508c38a",
# "build_date" : "2019-06-20T15:54:18.811730Z",
# "build_snapshot" : false,
# "lucene_version" : "8.0.0",
# "minimum_wire_compatibility_version" : "6.8.0",
# "minimum_index_compatibility_version" : "6.0.0-beta1"
# },
# "tagline" : "You Know, for Search"
# }
# verdi-es:~ #
#
curl --cacert /etc/elasticsearch/certs/http_ca.crt -u elastic https://localhost:${ELASTIC_SEARCH_PORT}
# ==========
# Install Kibana (kibana-oss)
# ==========
KIBANA_SERVER_HOST="localhost"
KIBANA_SERVER_PORT="5601"
# see: https://www.elastic.co/guide/en/kibana/current/deb.html
# Same as Elasticsearch, we will install the latest version of Kibana using the
# apt package manager from the official Elastic repository:
#
apt-get install kibana
# ---
# Adjust file '/etc/kibana/kibana.yml'
# ---
# see also: https://www.elastic.co/guide/en/kibana/8.3/settings.html
if [[ ! -f "/etc/kibana/kibana.yml.ORIG" ]] ; then
cp -a "/etc/kibana/kibana.yml" "/etc/kibana/kibana.yml.ORIG"
fi
# Specifies the address to which the Kibana server will bind. IP addresses and host
# names are both valid values. The default is 'localhost', which usually means
# remote machines will not be able to connect.
# To allow connections from remote users, set this parameter to a non-loopback address.
#
# Set server.host to 'localhost'
#
if ! grep -q -E "^\s*server.host:\s+${KIBANA_SERVER_HOST}" /etc/kibana/kibana.yml ; then
if grep -q -E "^\s*#server.host:" /etc/kibana/kibana.yml ; then
perl -i -n -p -e "s/^(#server.host:.*)/\1\nserver.host: ${KIBANA_SERVER_HOST}/" \
/etc/kibana/kibana.yml
else
cat << EOF >> /etc/kibana/kibana.yml
# Additional User Setting (server.host:)
#
server.host: ${KIBANA_SERVER_HOST}
EOF
fi
fi
# Set server.port to '5601'
#
if ! grep -q -E "^\s*server.port:\s+${KIBANA_SERVER_PORT}" /etc/kibana/kibana.yml ; then
if grep -q -E "^\s*#server.port:" /etc/kibana/kibana.yml ; then
perl -i -n -p -e "s/^(#server.port:.*)/\1\nserver.port: ${KIBANA_SERVER_PORT}/" \
/etc/kibana/kibana.yml
else
cat << EOF >> /etc/kibana/kibana.yml
# Additional User Setting (server.port:)
#
server.port: ${KIBANA_SERVER_HOST}
EOF
fi
fi
# ---
# Start Elasticsearch and generate an enrollment token for Kibana
# ---
# When you start Elasticsearch for the first time, the following security configuration
# occurs automatically:
#
# - Authentication and authorization are enabled, and a password is generated for
# the elastic built-in superuser.
#
# - Certificates and keys for TLS are generated for the transport and HTTP layer,
# and TLS is enabled and configured with these keys and certificates.
#
# The password and certificate and keys are output to your terminal.
#
# You can then generate an enrollment token for Kibana with the
# 'elasticsearch-create-enrollment-token tool:'
#
bin/elasticsearch-create-enrollment-token -s kibana
# ---
# Run Kibana with systemd
# ---
# Configure Kibana to start automatically when the system starts:
#
systemctl daemon-reload
systemctl enable kibana.service
# Kibana service ican be started and stopped as follows:
##
systemctl start kibana.service
systemctl istop kibana.service
# ==========
# Configure Nginx Reverse Proxy for Kibana
# ==========
# see: https://phoenixnap.com/kb/kibana-nginx-proxy
WEBSITE=oolm-shop-elk.oopen.de
cat <<EOF > /etc/nginx/sites-available/${WEBSITE}.conf
server {
listen 80;
listen [::]:80 ;
server_name ${WEBSITE};
return 301 https://\$host\$request_uri;
}
server {
listen 443 ssl http2;
listen [::]:443 ssl http2;
server_name ${WEBSITE};
root /var/www/html;
index index.html index.htm index.nginx-debian.html;
# Include location directive for Let's Encrypt ACME Challenge
#
# Needed for (automated) updating certificate
#
include snippets/letsencrypt-acme-challenge.conf;
# Diffie-Hellman parameter for DHE ciphersuites, recommended 2048 bits
#
# To generate a dhparam.pem file, run in a terminal
# openssl dhparam -out /etc/nginx/ssl/dhparam.pem 2048
#
ssl_dhparam /etc/nginx/ssl/dhparam.pem;
# Eable session resumption to improve https performance
ssl_session_cache shared:SSL:50m;
ssl_session_timeout 10m;
ssl_session_tickets off;
ssl_ecdh_curve secp384r1;
ssl_certificate /var/lib/dehydrated/certs/${WEBSITE}/fullchain.pem;
ssl_certificate_key /var/lib/dehydrated/certs/${WEBSITE}/privkey.pem;
ssl_protocols TLSv1.2 TLSv1.3;
# ECDHE better than DHE (faster) ECDHE & DHE GCM better than CBC (attacks on AES)
# Everything better than SHA1 (deprecated)
#
#ssl_ciphers 'ECDHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES256-SHA:ECDHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA'
ssl_ciphers ECDHE-RSA-AES256-GCM-SHA512:DHE-RSA-AES256-GCM-SHA512:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-SHA384;
ssl_prefer_server_ciphers on;
#resolver 192.168.42.129 8.8.8.8 valid=300s;
#resolver_timeout 5s;
add_header Strict-Transport-Security "max-age=31536000 always";
add_header X-Frame-Options DENY;
add_header X-Content-Type-Options nosniff;
add_header X-XSS-Protection "1; mode=block";
access_log /var/log/nginx/${WEBSITE}_access.log;
error_log /var/log/nginx/${WEBSITE}_error.log;
auth_basic "Authentication Required";
auth_basic_user_file /etc/nginx/htpasswd.kibana;
location / {
proxy_pass http://localhost:5601;
proxy_http_version 1.1;
proxy_set_header Upgrade \$http_upgrade;
proxy_set_header Connection 'upgrade';
proxy_set_header Host \$host;
proxy_cache_bypass \$http_upgrade;
}
}
EOF
# - Enable site ${WEBSITE}
# -
ln -s ../sites-available/${WEBSITE}.conf /etc/nginx/sites-enabled/
# - Create a basic authentication file with the openssl command:
# -
# - user: admin
# - password: $E%R&T/Z(U
# -
echo "admin:$(openssl passwd -apr1 '$E%R&T/Z(U')" | sudo tee -a /etc/nginx/htpasswd.kiba
# - Restart Nginx Webservice
# -
systemctl restart nginx
# ==========
# Install Logstash (logstash)
# ==========
# The final step is to install Logstash using the apt package manager from
# the official Elastic repository.
#
apt-get install logstash
# Start the Logstash service and set it to start automatically on boot:
#
systemctl restart logstash
systemctl enable logstash
# ---
# Note:
# The Logstash configuration depends on your personal preferences and the
# plugins you will use. You can find more information about how to configure
# Logstash here:
#
# https://www.elastic.co/guide/en/logstash/current/configuration.html
# ---