Compare commits
6 Commits
role_nomad
...
role_kuber
| Author | SHA1 | Date | |
|---|---|---|---|
| 7e83a43755 | |||
| 0d2e715980 | |||
| c899de75ae | |||
| 74ff4b483f | |||
| 5759e50510 | |||
| 62a4a67fad |
@@ -1,4 +0,0 @@
|
|||||||
netclient:
|
|
||||||
# Token to join default-network
|
|
||||||
# leave empty to ignore
|
|
||||||
join_network_token:
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
---
|
|
||||||
dependencies: []
|
|
||||||
#- role: docker
|
|
||||||
@@ -1,4 +0,0 @@
|
|||||||
- name: Deploy CA Certificate
|
|
||||||
ansible.builtin.copy:
|
|
||||||
src: secret_files/netmaker_server/ca/ca.crt
|
|
||||||
dest: /etc/ssl/certs/netmaker-ca.pem
|
|
||||||
@@ -1,25 +0,0 @@
|
|||||||
- name: Install Packages
|
|
||||||
# when: docker_file.stat.exists == False
|
|
||||||
package:
|
|
||||||
name:
|
|
||||||
- gpg
|
|
||||||
- gpg-agent
|
|
||||||
|
|
||||||
- name: Add netmaker-key
|
|
||||||
apt_key:
|
|
||||||
url: https://apt.netmaker.org/gpg.key
|
|
||||||
state: present
|
|
||||||
|
|
||||||
- name: Add netmaker-repository
|
|
||||||
apt_repository:
|
|
||||||
repo: "deb https:apt.netmaker.org stable main"
|
|
||||||
state: present
|
|
||||||
filename: netmaker
|
|
||||||
update_cache: yes
|
|
||||||
|
|
||||||
- name: Install wireguard & netclient
|
|
||||||
package:
|
|
||||||
name:
|
|
||||||
- wireguard
|
|
||||||
- netclient
|
|
||||||
state: latest
|
|
||||||
@@ -1,7 +0,0 @@
|
|||||||
- name: Join netmaker-network
|
|
||||||
when: "netclient.join_network_token is defined"
|
|
||||||
command: "netclient join -t {{ netclient.join_network_token }}"
|
|
||||||
failed_when: command.rc != 0
|
|
||||||
changed_when: "'starting wireguard' in command.stdout"
|
|
||||||
register: command
|
|
||||||
throttle: 1
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
- import_tasks: ./certs.yml
|
|
||||||
|
|
||||||
- import_tasks: ./install.yml
|
|
||||||
|
|
||||||
- import_tasks: ./join-network.yml
|
|
||||||
|
|
||||||
- name: Gather facts to get changes
|
|
||||||
ansible.builtin.gather_facts:
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
netmaker_creds:
|
|
||||||
rqlite_password:
|
|
||||||
mq_admin_password:
|
|
||||||
|
|
||||||
master_key:
|
|
||||||
@@ -1,28 +0,0 @@
|
|||||||
# Overwrite for specific nodes to force dynamic-ip (disable setting public-ip and forces external lookup for public-ip)
|
|
||||||
# When false, will check itself for dynamic-ip (based on private-ip)
|
|
||||||
netmaker_dynamicIp: false
|
|
||||||
|
|
||||||
netmaker_nginx:
|
|
||||||
# Listen-port
|
|
||||||
tls_port: 51820
|
|
||||||
# Advertise-Port for services
|
|
||||||
# (must also be reachable by internal services!)
|
|
||||||
advertise_port: 51820
|
|
||||||
|
|
||||||
# This is the base-domain used for generating hostnames for services
|
|
||||||
netmaker_base_domain:
|
|
||||||
|
|
||||||
# host + base_domain
|
|
||||||
netmaker_api:
|
|
||||||
host: netmaker-api
|
|
||||||
netmaker_ui:
|
|
||||||
host: netmaker-ui
|
|
||||||
# MQTT-broker
|
|
||||||
netmaker_broker:
|
|
||||||
tls_host: netmaker-broker
|
|
||||||
|
|
||||||
# host + node_hostname
|
|
||||||
netmaker_rqlite:
|
|
||||||
http_host: netmaker-rqlite-http
|
|
||||||
cluster_host: netmaker-rqlite-cluster
|
|
||||||
|
|
||||||
@@ -1,37 +0,0 @@
|
|||||||
@startuml
|
|
||||||
|
|
||||||
interface ng_TLS
|
|
||||||
|
|
||||||
component netmaker_server {
|
|
||||||
component nginx {
|
|
||||||
component ng_stream
|
|
||||||
component ng_http
|
|
||||||
|
|
||||||
ng_stream -up- ng_TLS
|
|
||||||
|
|
||||||
ng_stream -right-> ng_http : tls-termination
|
|
||||||
}
|
|
||||||
|
|
||||||
component nm_ui
|
|
||||||
nm_ui -up- nm_ui_http
|
|
||||||
ng_http -down-( nm_ui_http
|
|
||||||
|
|
||||||
component Mosquitto
|
|
||||||
Mosquitto -up- mq_plain
|
|
||||||
Mosquitto -up- mq_tls
|
|
||||||
ng_stream -down-( mq_tls
|
|
||||||
|
|
||||||
component rqlite
|
|
||||||
rqlite -up- rq_http
|
|
||||||
rqlite -up- rq_cluster
|
|
||||||
ng_stream -down-( rq_cluster
|
|
||||||
ng_http -down-( rq_http
|
|
||||||
|
|
||||||
component nm_api
|
|
||||||
nm_api -down- nm_api_http
|
|
||||||
ng_http --( nm_api_http
|
|
||||||
nm_api .up.( ng_TLS : db-connection to rqlite-master
|
|
||||||
nm_api --( mq_plain
|
|
||||||
}
|
|
||||||
|
|
||||||
@enduml
|
|
||||||
@@ -1,12 +0,0 @@
|
|||||||
per_listener_settings false
|
|
||||||
|
|
||||||
listener 8883
|
|
||||||
protocol websockets
|
|
||||||
allow_anonymous false
|
|
||||||
|
|
||||||
listener 1883
|
|
||||||
protocol websockets
|
|
||||||
allow_anonymous false
|
|
||||||
|
|
||||||
plugin /usr/lib/mosquitto_dynamic_security.so
|
|
||||||
plugin_opt_config_file /mosquitto/data/dynamic-security.json
|
|
||||||
@@ -1,23 +0,0 @@
|
|||||||
#!/bin/ash
|
|
||||||
|
|
||||||
wait_for_netmaker() {
|
|
||||||
echo "SERVER: ${NETMAKER_SERVER_HOST}"
|
|
||||||
until curl --output /dev/null --silent --fail --head \
|
|
||||||
--location "${NETMAKER_SERVER_HOST}/api/server/health"; do
|
|
||||||
echo "Waiting for netmaker server to startup"
|
|
||||||
sleep 1
|
|
||||||
done
|
|
||||||
}
|
|
||||||
|
|
||||||
main(){
|
|
||||||
# wait for netmaker to startup
|
|
||||||
apk add curl
|
|
||||||
wait_for_netmaker
|
|
||||||
echo "Starting MQ..."
|
|
||||||
# Run the main container command.
|
|
||||||
/docker-entrypoint.sh
|
|
||||||
/usr/sbin/mosquitto -c /mosquitto/config/mosquitto.conf
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
main "${@}"
|
|
||||||
@@ -1,33 +0,0 @@
|
|||||||
|
|
||||||
user nginx;
|
|
||||||
worker_processes auto;
|
|
||||||
|
|
||||||
error_log /var/log/nginx/error.log notice;
|
|
||||||
pid /var/run/nginx.pid;
|
|
||||||
|
|
||||||
|
|
||||||
events {
|
|
||||||
worker_connections 1024;
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
http {
|
|
||||||
include /etc/nginx/mime.types;
|
|
||||||
default_type application/octet-stream;
|
|
||||||
|
|
||||||
log_format main '$remote_addr - $remote_user [$time_local] "$request" '
|
|
||||||
'$status $body_bytes_sent "$http_referer" '
|
|
||||||
'"$http_user_agent" "$http_x_forwarded_for"';
|
|
||||||
|
|
||||||
access_log /var/log/nginx/access.log main;
|
|
||||||
|
|
||||||
sendfile on;
|
|
||||||
#tcp_nopush on;
|
|
||||||
|
|
||||||
keepalive_timeout 65;
|
|
||||||
|
|
||||||
#gzip on;
|
|
||||||
|
|
||||||
include /etc/nginx/conf.d/*.conf;
|
|
||||||
}
|
|
||||||
include /etc/nginx/stream.d/*.conf;
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
---
|
|
||||||
# dependencies:
|
|
||||||
# - role: docker
|
|
||||||
@@ -1,40 +0,0 @@
|
|||||||
- name: Generate PrivateKey
|
|
||||||
community.crypto.openssl_privatekey:
|
|
||||||
path: /opt/netmaker_server/certs/node.key
|
|
||||||
owner: 1883 # Set owner to mosquitto-user (all other containers seem to run as root)
|
|
||||||
|
|
||||||
- name: Generate Certificate-Signing-Request from privateKey
|
|
||||||
community.crypto.openssl_csr:
|
|
||||||
path: /opt/netmaker_server/certs/node.csr
|
|
||||||
privatekey_path: /opt/netmaker_server/certs/node.key
|
|
||||||
common_name: "{{ ansible_facts.nodename }}"
|
|
||||||
subject_alt_name:
|
|
||||||
"DNS:{{ netmaker_rqlite.http_host }}.{{ ansible_facts.nodename }},\
|
|
||||||
DNS:{{ netmaker_rqlite.cluster_host }}.{{ ansible_facts.nodename }},\
|
|
||||||
DNS:{{ netmaker_broker.tls_host }}.{{ netmaker_base_domain }},\
|
|
||||||
DNS:{{ netmaker_api.host }}.{{ netmaker_base_domain }},\
|
|
||||||
DNS:{{ netmaker_ui.host }}.{{ netmaker_base_domain }}"
|
|
||||||
|
|
||||||
- name: Fetch CSR
|
|
||||||
ansible.builtin.fetch:
|
|
||||||
src: /opt/netmaker_server/certs/node.csr
|
|
||||||
dest: tmp_files/
|
|
||||||
|
|
||||||
- name: Sign CSR locally with CA
|
|
||||||
local_action: community.crypto.x509_certificate
|
|
||||||
args:
|
|
||||||
path: tmp_files/{{ inventory_hostname }}/opt/netmaker_server/certs/node.crt
|
|
||||||
csr_path: tmp_files/{{ inventory_hostname }}/opt/netmaker_server/certs/node.csr
|
|
||||||
ownca_path: secret_files/netmaker_server/ca/ca.crt
|
|
||||||
ownca_privatekey_path: secret_files/netmaker_server/ca/ca.key
|
|
||||||
provider: ownca
|
|
||||||
|
|
||||||
- name: Copy Signed Certificate
|
|
||||||
ansible.builtin.copy:
|
|
||||||
src: tmp_files/{{ inventory_hostname }}/opt/netmaker_server/certs/node.crt
|
|
||||||
dest: /opt/netmaker_server/certs/node.crt
|
|
||||||
|
|
||||||
- name: Copy CA Certificate
|
|
||||||
ansible.builtin.copy:
|
|
||||||
src: secret_files/netmaker_server/ca/ca.crt
|
|
||||||
dest: /opt/netmaker_server/certs/ca.crt
|
|
||||||
@@ -1,20 +0,0 @@
|
|||||||
- import_tasks: ./prerequisites.yml
|
|
||||||
|
|
||||||
- name: Copy folder-structure
|
|
||||||
ansible.builtin.copy:
|
|
||||||
src: opt/netmaker_server
|
|
||||||
dest: /opt/
|
|
||||||
mode: preserve
|
|
||||||
|
|
||||||
- name: Deploy compose file
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: docker-compose.yml.template
|
|
||||||
dest: /opt/netmaker_server/docker-compose.yml
|
|
||||||
|
|
||||||
- import_tasks: ./certs.yml
|
|
||||||
|
|
||||||
- import_tasks: ./nginx.yml
|
|
||||||
|
|
||||||
- import_tasks: ./rqlite.yml
|
|
||||||
|
|
||||||
- import_tasks: ./netmaker.yml
|
|
||||||
@@ -1,57 +0,0 @@
|
|||||||
- name: Start rest of netmaker-services
|
|
||||||
command: "docker-compose --project-directory /opt/netmaker_server/ up -d"
|
|
||||||
register: command
|
|
||||||
failed_when: command.rc != 0
|
|
||||||
|
|
||||||
- name: Wait for netmaker-api to become available
|
|
||||||
uri:
|
|
||||||
url: 'https://{{ netmaker_api.host }}.{{ netmaker_base_domain }}:{{ netmaker_nginx.advertise_port }}'
|
|
||||||
return_content: yes
|
|
||||||
validate_certs: no
|
|
||||||
status_code:
|
|
||||||
- 404
|
|
||||||
until: uri_output.status == 404
|
|
||||||
retries: 24 # Retries for 24 * 5 seconds = 120 seconds = 2 minutes
|
|
||||||
delay: 5 # Every 5 seconds
|
|
||||||
register: uri_output
|
|
||||||
|
|
||||||
# todo: check if exists?
|
|
||||||
|
|
||||||
- name: Create default mesh-network 'server'
|
|
||||||
uri:
|
|
||||||
validate_certs: no
|
|
||||||
url: 'https://{{ netmaker_api.host }}.{{ netmaker_base_domain }}:{{ netmaker_nginx.advertise_port }}/api/networks'
|
|
||||||
method: POST
|
|
||||||
body:
|
|
||||||
netid: servnet
|
|
||||||
addressrange: 10.92.0.0/24
|
|
||||||
addressrange6: fd92::/64
|
|
||||||
body_format: json
|
|
||||||
headers:
|
|
||||||
Authorization: 'Bearer {{ netmaker_creds.master_key }}'
|
|
||||||
Content-Type: application/json
|
|
||||||
when: "inventory_hostname == groups['netmaker_server'][0]"
|
|
||||||
register: default_mesh
|
|
||||||
until: "default_mesh is not failed"
|
|
||||||
retries: 2
|
|
||||||
delay: 10
|
|
||||||
|
|
||||||
# todo: check if exists?
|
|
||||||
|
|
||||||
- name: Create token for default-network
|
|
||||||
uri:
|
|
||||||
validate_certs: no
|
|
||||||
url: 'https://{{ netmaker_api.host }}.{{ netmaker_base_domain }}:{{ netmaker_nginx.advertise_port }}/api/networks/servnet/keys' # todo: do implementation
|
|
||||||
method: POST
|
|
||||||
body:
|
|
||||||
name: ""
|
|
||||||
uses: 0
|
|
||||||
body_format: json
|
|
||||||
headers:
|
|
||||||
Authorization: 'Bearer {{ netmaker_creds.master_key }}'
|
|
||||||
Content-Type: application/json
|
|
||||||
when: "inventory_hostname == groups['netmaker_server'][0]"
|
|
||||||
register: default_mesh_key
|
|
||||||
until: "default_mesh_key is not failed"
|
|
||||||
retries: 2
|
|
||||||
delay: 10
|
|
||||||
@@ -1,18 +0,0 @@
|
|||||||
- name: Deploy nginx configs
|
|
||||||
template:
|
|
||||||
src: "{{item.src}}"
|
|
||||||
dest: "{{item.dst}}"
|
|
||||||
loop:
|
|
||||||
- { src: 'nginx/proxy.conf.template', dst: '/opt/netmaker_server/nginx/conf/conf.d/proxy.conf' }
|
|
||||||
- { src: 'nginx/passthrough.conf.template', dst: '/opt/netmaker_server/nginx/conf/stream.d/passthrough.conf' }
|
|
||||||
|
|
||||||
- name: Start nginx service
|
|
||||||
command: "docker-compose --project-directory /opt/netmaker_server/ up -d nginx"
|
|
||||||
register: command
|
|
||||||
failed_when: command.rc != 0
|
|
||||||
|
|
||||||
- name: Waiting for nginx to accept connections
|
|
||||||
ansible.builtin.wait_for:
|
|
||||||
host: "{{ inventory_hostname }}"
|
|
||||||
port: 51820
|
|
||||||
state: started
|
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
- name: Install wireguard
|
|
||||||
package:
|
|
||||||
name:
|
|
||||||
- wireguard
|
|
||||||
state: latest
|
|
||||||
|
|
||||||
- name: Check if default-ipv4-address is private
|
|
||||||
set_fact:
|
|
||||||
private_ipv4_address: "{{ ansible_facts.default_ipv4.address | regex_search('^((10)|(192\\.168)|(172\\.((1[6-9])|(2[0-9])|(3[0-1])))|(100))\\.') }}"
|
|
||||||
@@ -1,42 +0,0 @@
|
|||||||
- name: Deploy rqlite config
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: rqlite-config.json.template
|
|
||||||
dest: /opt/netmaker_server/rqlite/config.json
|
|
||||||
|
|
||||||
- name: Start rqlite service for 1st-node
|
|
||||||
command: "docker-compose --project-directory /opt/netmaker_server/ up -d rqlite"
|
|
||||||
register: command
|
|
||||||
failed_when: command.rc != 0
|
|
||||||
when: "inventory_hostname == groups['netmaker_server'][0]"
|
|
||||||
|
|
||||||
- name: Waiting for rqlite to accept connections on 1st-node
|
|
||||||
uri:
|
|
||||||
url: 'https://{{ netmaker_rqlite.http_host }}.{{ inventory_hostname }}:{{ netmaker_nginx.advertise_port }}/status'
|
|
||||||
return_content: yes
|
|
||||||
validate_certs: no
|
|
||||||
status_code:
|
|
||||||
- 401
|
|
||||||
until: uri_output.status == 401
|
|
||||||
retries: 24 # Retries for 24 * 5 seconds = 120 seconds = 2 minutes
|
|
||||||
delay: 5 # Every 5 seconds
|
|
||||||
register: uri_output
|
|
||||||
when: "inventory_hostname == groups['netmaker_server'][0]"
|
|
||||||
|
|
||||||
- name: Start rqlite service for other nodes
|
|
||||||
command: "docker-compose --project-directory /opt/netmaker_server/ up -d rqlite"
|
|
||||||
register: command
|
|
||||||
failed_when: command.rc != 0
|
|
||||||
when: "inventory_hostname != groups['netmaker_server'][0]"
|
|
||||||
|
|
||||||
- name: Waiting for rqlite to accept connections on other nodes
|
|
||||||
uri:
|
|
||||||
url: 'https://{{ netmaker_rqlite.http_host }}.{{ inventory_hostname }}:{{ netmaker_nginx.advertise_port }}/status'
|
|
||||||
return_content: yes
|
|
||||||
validate_certs: no
|
|
||||||
status_code:
|
|
||||||
- 401
|
|
||||||
until: uri_output.status == 401
|
|
||||||
retries: 24 # Retries for 24 * 5 seconds = 120 seconds = 2 minutes
|
|
||||||
delay: 5 # Every 5 seconds
|
|
||||||
register: uri_output
|
|
||||||
when: "inventory_hostname != groups['netmaker_server'][0]"
|
|
||||||
@@ -1,127 +0,0 @@
|
|||||||
version: "3.4"
|
|
||||||
|
|
||||||
services:
|
|
||||||
nginx:
|
|
||||||
image: nginx
|
|
||||||
restart: unless-stopped
|
|
||||||
volumes:
|
|
||||||
- ./nginx/nginx.conf:/etc/nginx/nginx.conf:ro # Override nginx-config to add stream-import
|
|
||||||
- ./nginx/conf/conf.d/:/etc/nginx/conf.d:ro # conf.d
|
|
||||||
- ./nginx/conf/stream.d/:/etc/nginx/stream.d:ro # conf.d
|
|
||||||
- ./certs:/certs:ro # SSL-certificates
|
|
||||||
ports:
|
|
||||||
- {{ netmaker_nginx.tls_port }}:443
|
|
||||||
|
|
||||||
rqlite: # Distributed sqlite-db
|
|
||||||
image: rqlite/rqlite
|
|
||||||
restart: unless-stopped
|
|
||||||
hostname: "{{ ansible_facts.nodename }}"
|
|
||||||
volumes:
|
|
||||||
- "./rqlite/data:/rqlite/file"
|
|
||||||
- "./rqlite/config.json:/config.json:ro"
|
|
||||||
- "./certs:/certs:ro"
|
|
||||||
- ./certs/ca.crt:/etc/ssl/certs/netmaker-ca.pem:ro # Add CA to system-trust-store
|
|
||||||
command: "
|
|
||||||
-http-adv-addr {{ netmaker_rqlite.http_host }}.{{ ansible_facts.nodename }}:{{ netmaker_nginx.advertise_port }}
|
|
||||||
-raft-addr [::]:4002
|
|
||||||
-raft-adv-addr {{ netmaker_rqlite.cluster_host }}.{{ ansible_facts.nodename }}:{{ netmaker_nginx.advertise_port }}
|
|
||||||
|
|
||||||
-node-encrypt
|
|
||||||
-node-cert /certs/node.crt
|
|
||||||
-node-key /certs/node.key
|
|
||||||
-node-no-verify
|
|
||||||
|
|
||||||
-auth /config.json
|
|
||||||
|
|
||||||
{% if inventory_hostname != groups['netmaker_server'][0] %}
|
|
||||||
-join-as netmaker
|
|
||||||
-join https://{{ netmaker_rqlite.http_host }}.{{ groups['netmaker_server'][0] }}:{{ netmaker_nginx.advertise_port }}
|
|
||||||
{% endif %}
|
|
||||||
"
|
|
||||||
# FIXME: /\ \/ Change http -> https
|
|
||||||
|
|
||||||
netmaker: # The Primary Server for running Netmaker
|
|
||||||
image: gravitl/netmaker:v0.17.1
|
|
||||||
depends_on:
|
|
||||||
- rqlite
|
|
||||||
cap_add:
|
|
||||||
- NET_ADMIN
|
|
||||||
- NET_RAW
|
|
||||||
- SYS_MODULE
|
|
||||||
sysctls:
|
|
||||||
- net.ipv4.ip_forward=1
|
|
||||||
- net.ipv4.conf.all.src_valid_mark=1
|
|
||||||
- net.ipv6.conf.all.disable_ipv6=0
|
|
||||||
- net.ipv6.conf.all.forwarding=1
|
|
||||||
restart: unless-stopped
|
|
||||||
volumes: # Volume mounts necessary for sql, coredns, and mqtt
|
|
||||||
- ./dnsconfig/:/root/config/dnsconfig
|
|
||||||
- ./mosquitto/data/:/etc/netmaker/
|
|
||||||
- ./certs/ca.crt:/etc/ssl/certs/netmaker-ca.pem:ro # Add CA to system-trust-store
|
|
||||||
hostname: "{{ ansible_facts.nodename }}"
|
|
||||||
environment: # Necessary capabilities to set iptables when running in container
|
|
||||||
NODE_ID: "{{ ansible_facts.nodename }}"
|
|
||||||
MASTER_KEY: "{{ netmaker_creds.master_key }}" # The admin master key for accessing the API. Change this in any production installation.
|
|
||||||
|
|
||||||
{% if not private_ipv4_address and not netmaker_dynamicIp %}
|
|
||||||
SERVER_HOST: "{{ ansible_facts.default_ipv4.address }}" # Set to public IP of machine.
|
|
||||||
{% endif %}
|
|
||||||
SERVER_NAME: "{{ netmaker_broker.tls_host }}.{{ netmaker_base_domain }}" # The domain/host IP indicating the mq broker address
|
|
||||||
SERVER_HTTP_HOST: "{{ netmaker_api.host }}.{{ netmaker_base_domain }}" # Overrides SERVER_HOST if set. Useful for making HTTP available via different interfaces/networks.
|
|
||||||
SERVER_API_CONN_STRING: "{{ netmaker_api.host }}.{{ netmaker_base_domain }}:{{ netmaker_nginx.advertise_port }}"
|
|
||||||
|
|
||||||
DISABLE_REMOTE_IP_CHECK: "off" # If turned "on", Server will not set Host based on remote IP check. This is already overridden if SERVER_HOST is set. Turned "off" by default.
|
|
||||||
DNS_MODE: "off" # Enables DNS Mode, meaning all nodes will set hosts file for private dns settings.
|
|
||||||
|
|
||||||
API_PORT: "8081" # The HTTP API port for Netmaker. Used for API calls / communication from front end. If changed, need to change port of BACKEND_URL for netmaker-ui.
|
|
||||||
REST_BACKEND: "on" # Enables the REST backend (API running on API_PORT at SERVER_HTTP_HOST). Change to "off" to turn off.
|
|
||||||
RCE: "off" # Enables setting PostUp and PostDown (arbitrary commands) on nodes from the server. Off by default.
|
|
||||||
CORS_ALLOWED_ORIGIN: "*" # The "allowed origin" for API requests. Change to restrict where API requests can come from.
|
|
||||||
DISPLAY_KEYS: "on" # Show keys permanently in UI (until deleted) as opposed to 1-time display.
|
|
||||||
|
|
||||||
DATABASE: "rqlite"
|
|
||||||
SQL_CONN: "https://netmaker:{{ netmaker_creds.rqlite_password }}@{{ netmaker_rqlite.http_host }}.{{ ansible_facts.nodename }}:{{ netmaker_nginx.advertise_port }}/"
|
|
||||||
|
|
||||||
MQ_HOST: "mosquitto" # the address of the mq server. If running from docker compose it will be "mq". Otherwise, need to input address. If using "host networking", it will find and detect the IP of the mq container.
|
|
||||||
MQ_SERVER_PORT: "1883" # the reachable port of MQ by the server - change if internal MQ port changes (or use external port if MQ is not on the same machine)
|
|
||||||
MQ_PORT: "{{ netmaker_nginx.advertise_port }}" # the reachable port of MQ - change if external MQ port changes (port on proxy, not necessarily the one exposed in docker-compose)
|
|
||||||
MQ_ADMIN_PASSWORD: "{{ netmaker_creds.mq_admin_password }}"
|
|
||||||
|
|
||||||
HOST_NETWORK: "off" # whether or not host networking is turned on. Only turn on if configured for host networking (see docker-compose.hostnetwork.yml). Will set host-level settings like iptables.
|
|
||||||
PORT_FORWARD_SERVICES: "" # decide which services to port forward ("dns","ssh", or "mq")
|
|
||||||
|
|
||||||
# this section is for OAuth
|
|
||||||
AUTH_PROVIDER: "" # "<azure-ad|github|google|oidc>"
|
|
||||||
CLIENT_ID: "" # "<client id of your oauth provider>"
|
|
||||||
CLIENT_SECRET: "" # "<client secret of your oauth provider>"
|
|
||||||
FRONTEND_URL: "" # "https://dashboard.<netmaker base domain>"
|
|
||||||
AZURE_TENANT: "" # "<only for azure, you may optionally specify the tenant for the OAuth>"
|
|
||||||
OIDC_ISSUER: "" # https://oidc.yourprovider.com - URL of oidc provider
|
|
||||||
|
|
||||||
VERBOSITY: "1" # logging verbosity level - 1, 2, or 3
|
|
||||||
TELEMETRY: "off" # Whether or not to send telemetry data to help improve Netmaker. Switch to "off" to opt out of sending telemetry.
|
|
||||||
ports:
|
|
||||||
- "51821-51830:51821-51830/udp" # wireguard ports
|
|
||||||
|
|
||||||
netmaker-ui: # The Netmaker UI Component
|
|
||||||
image: gravitl/netmaker-ui:v0.17.1
|
|
||||||
depends_on:
|
|
||||||
- netmaker
|
|
||||||
links:
|
|
||||||
- "netmaker:api"
|
|
||||||
restart: unless-stopped
|
|
||||||
environment:
|
|
||||||
BACKEND_URL: "https://{{ netmaker_api.host }}.{{ netmaker_base_domain }}:{{ netmaker_nginx.advertise_port }}" # URL where UI will send API requests. Change based on SERVER_HOST, SERVER_HTTP_HOST, and API_PORT
|
|
||||||
|
|
||||||
mosquitto: # the MQTT broker for netmaker
|
|
||||||
image: eclipse-mosquitto:2.0.11-openssl
|
|
||||||
restart: unless-stopped
|
|
||||||
volumes:
|
|
||||||
- ./mosquitto/config:/mosquitto/config
|
|
||||||
- ./mosquitto/data:/mosquitto/data
|
|
||||||
- ./mosquitto/logs:/mosquitto/log
|
|
||||||
depends_on:
|
|
||||||
- netmaker
|
|
||||||
command: ["/mosquitto/config/wait.sh"]
|
|
||||||
environment:
|
|
||||||
NETMAKER_SERVER_HOST: "http://netmaker:8081"
|
|
||||||
@@ -1,25 +0,0 @@
|
|||||||
stream{
|
|
||||||
# Map target-hosts based on hostname
|
|
||||||
map $ssl_preread_server_name $target_host {
|
|
||||||
hostnames; # Enable matching including prefix/suffix-mask
|
|
||||||
|
|
||||||
{{ netmaker_ui.host }}.{{ netmaker_base_domain }} 127.0.0.1:8443;
|
|
||||||
{{ netmaker_api.host }}.{{ netmaker_base_domain }} 127.0.0.1:8443;
|
|
||||||
|
|
||||||
{{ netmaker_broker.tls_host }}.{{ netmaker_base_domain }} 127.0.0.1:8443;
|
|
||||||
|
|
||||||
{{ netmaker_rqlite.http_host }}.{{ ansible_facts.nodename }} 127.0.0.1:8443;
|
|
||||||
{{ netmaker_rqlite.cluster_host }}.{{ ansible_facts.nodename }} rqlite:4002;
|
|
||||||
|
|
||||||
default 127.0.0.1:1;
|
|
||||||
}
|
|
||||||
|
|
||||||
server {
|
|
||||||
resolver 127.0.0.11; # Explicitly set docker-resolver
|
|
||||||
|
|
||||||
listen 443;
|
|
||||||
ssl_preread on;
|
|
||||||
|
|
||||||
proxy_pass $target_host;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,29 +0,0 @@
|
|||||||
map $host $proxy_name {
|
|
||||||
hostnames;
|
|
||||||
|
|
||||||
{{ netmaker_ui.host }}.{{ netmaker_base_domain }} netmaker-ui:80;
|
|
||||||
{{ netmaker_api.host }}.{{ netmaker_base_domain }} netmaker:8081;
|
|
||||||
|
|
||||||
{{ netmaker_broker.tls_host }}.{{ netmaker_base_domain }} mosquitto:8883;
|
|
||||||
|
|
||||||
{{ netmaker_rqlite.http_host }}.{{ ansible_facts.nodename }} rqlite:4001;
|
|
||||||
|
|
||||||
default 444;
|
|
||||||
}
|
|
||||||
|
|
||||||
server {
|
|
||||||
resolver 127.0.0.11; # Explicitly set docker-resolver
|
|
||||||
|
|
||||||
listen 8443 ssl;
|
|
||||||
|
|
||||||
ssl_certificate /certs/node.crt;
|
|
||||||
ssl_certificate_key /certs/node.key;
|
|
||||||
|
|
||||||
if ($proxy_name = 444){
|
|
||||||
return 444;
|
|
||||||
}
|
|
||||||
|
|
||||||
location / {
|
|
||||||
proxy_pass http://$proxy_name;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
[{
|
|
||||||
"username": "netmaker",
|
|
||||||
"password": "{{ netmaker_creds.rqlite_password }}",
|
|
||||||
"perms": ["all"]
|
|
||||||
}]
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
ipv6_stable_secret: 1111:2222:3333:4444:5555:6666:7777:8888
|
|
||||||
@@ -5,5 +5,3 @@
|
|||||||
- import_tasks: ./packages.yml
|
- import_tasks: ./packages.yml
|
||||||
|
|
||||||
- import_tasks: ./aliases.yml
|
- import_tasks: ./aliases.yml
|
||||||
|
|
||||||
- import_tasks: ./networking.yml
|
|
||||||
|
|||||||
@@ -1,22 +0,0 @@
|
|||||||
- name: Set sysctl settings for ip-forwarding
|
|
||||||
copy:
|
|
||||||
dest: "/etc/sysctl.d/ip-forwarding.conf"
|
|
||||||
content: |
|
|
||||||
net.ipv4.ip_forward = 1
|
|
||||||
net.ipv6.conf.all.forwarding = 1
|
|
||||||
notify: reload_sysctl
|
|
||||||
|
|
||||||
- name: Set sysctl settings for ipv6-address-generation
|
|
||||||
copy:
|
|
||||||
dest: "/etc/sysctl.d/ipv6-slaac-address-generation.conf"
|
|
||||||
content: |
|
|
||||||
net.ipv6.conf.default.addr_gen_mode = 2
|
|
||||||
net.ipv6.conf.default.stable_secret = {{ ipv6_stable_secret }}
|
|
||||||
notify: reload_sysctl
|
|
||||||
|
|
||||||
- name: Set sysctl settings to override ipv6-slaac with enabled forwarding
|
|
||||||
copy:
|
|
||||||
dest: "/etc/sysctl.d/ipv6-slaac-override.conf"
|
|
||||||
content: |
|
|
||||||
net.ipv6.conf.all.accept_ra = 2
|
|
||||||
notify: reload_sysctl
|
|
||||||
@@ -1,17 +0,0 @@
|
|||||||
---
|
|
||||||
kubernetes:
|
|
||||||
ipPool:
|
|
||||||
ipv4:
|
|
||||||
cluster_cidr: 10.42.0.0/16
|
|
||||||
service_cidr: 10.43.0.0/16
|
|
||||||
ipv6:
|
|
||||||
cluster_cidr: fd42::/56
|
|
||||||
service_cidr: fd43::/112
|
|
||||||
|
|
||||||
# Replace - with _
|
|
||||||
nodeIp_interface: <interface to grab nodeIp from>
|
|
||||||
|
|
||||||
control_plane:
|
|
||||||
dns_name: <control-plane dns-reachable-name>
|
|
||||||
|
|
||||||
token: <shared token for nodes to join>
|
|
||||||
@@ -1,33 +0,0 @@
|
|||||||
@startuml
|
|
||||||
|
|
||||||
rectangle "Control-Plane" as control_plane {
|
|
||||||
rectangle "Node" as sn1 {
|
|
||||||
component "netclient" as sn1_netclient
|
|
||||||
|
|
||||||
component etcd as sn1_etcd
|
|
||||||
component "k3s-server" as sn1_k3s_server
|
|
||||||
sn1_k3s_server - sn1_etcd
|
|
||||||
}
|
|
||||||
|
|
||||||
rectangle "Node" as sn2 {
|
|
||||||
component "netclient" as sn2_netclient
|
|
||||||
|
|
||||||
component etcd as sn2_etcd
|
|
||||||
component "k3s-server" as sn2_k3s_server
|
|
||||||
sn2_k3s_server - sn2_etcd
|
|
||||||
}
|
|
||||||
|
|
||||||
sn1_netclient -- sn2_netclient
|
|
||||||
sn1_etcd -- sn2_etcd
|
|
||||||
}
|
|
||||||
|
|
||||||
rectangle "Workers" {
|
|
||||||
rectangle "Node" as an1 {
|
|
||||||
component "netclient" as an1_netclient
|
|
||||||
|
|
||||||
component "k3s-agent" as sn1_k3s_agent
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
@enduml
|
|
||||||
@@ -31,5 +31,6 @@ disabled_plugins = []
|
|||||||
# level = "info"
|
# level = "info"
|
||||||
|
|
||||||
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc]
|
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc]
|
||||||
|
|
||||||
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc.options]
|
[plugins."io.containerd.grpc.v1.cri".containerd.runtimes.runc.options]
|
||||||
SystemdCgroup = true
|
SystemdCgroup = true
|
||||||
|
|||||||
15
kubernetes/files/k8s-dqlite/Dockerfile
Normal file
15
kubernetes/files/k8s-dqlite/Dockerfile
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
FROM golang:1-buster
|
||||||
|
|
||||||
|
# Add PPA
|
||||||
|
RUN echo "deb http://ppa.launchpad.net/dqlite/dev/ubuntu bionic main" > /etc/apt/sources.list.d/ppa_dqlite_dev_bionic.list
|
||||||
|
RUN apt-key adv --keyserver keyserver.ubuntu.com --recv-keys 50FB3D04
|
||||||
|
# Install dependencies
|
||||||
|
RUN apt update -y && apt install -y build-essential git libraft-dev libsqlite3-dev libdqlite-dev
|
||||||
|
|
||||||
|
# Clone
|
||||||
|
RUN git clone https://github.com/canonical/k8s-dqlite --branch v1.0.4 /k8s-dqlite
|
||||||
|
WORKDIR /k8s-dqlite
|
||||||
|
|
||||||
|
# Compile
|
||||||
|
ENV CGO_LDFLAGS_ALLOW="-Wl,-z,now"
|
||||||
|
RUN go build -o k8s-dqlite -tags libsqlite3,dqlite k8s-dqlite.go
|
||||||
BIN
kubernetes/files/k8s-dqlite/k8s-dqlite
Executable file
BIN
kubernetes/files/k8s-dqlite/k8s-dqlite
Executable file
Binary file not shown.
@@ -1,19 +1,3 @@
|
|||||||
- name: reload_sysctl
|
- name: reload_sysctl
|
||||||
command: sysctl --system
|
command: sysctl --system
|
||||||
|
|
||||||
- name: restart_containerd
|
|
||||||
ansible.builtin.service:
|
|
||||||
name: containerd
|
|
||||||
state: restarted
|
|
||||||
|
|
||||||
- name: reload_networking
|
|
||||||
service:
|
|
||||||
name: networking
|
|
||||||
state: restarted
|
|
||||||
async: 5
|
|
||||||
poll: 0
|
|
||||||
notify: wait_for_connection
|
|
||||||
|
|
||||||
- name: wait_for_connection
|
|
||||||
wait_for_connection:
|
|
||||||
delay: 5
|
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
---
|
---
|
||||||
dependencies:
|
dependencies:
|
||||||
- role: docker
|
- { role: docker }
|
||||||
- role: netmaker
|
|
||||||
|
|||||||
12
kubernetes/tasks/deploy_cilium.yml
Normal file
12
kubernetes/tasks/deploy_cilium.yml
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
- name: Deploy Cilium-CLI
|
||||||
|
ansible.builtin.unarchive:
|
||||||
|
src: https://github.com/cilium/cilium-cli/releases/latest/download/cilium-linux-amd64.tar.gz
|
||||||
|
dest: /usr/local/bin
|
||||||
|
remote_src: yes
|
||||||
|
mode: u=rwx,g=rx,o=rx
|
||||||
|
|
||||||
|
- name: Install Cilium
|
||||||
|
when: "inventory_hostname == groups['kubernetes'][0]"
|
||||||
|
command: -cilium install
|
||||||
|
environment:
|
||||||
|
KUBECONFIG: /etc/kubernetes/admin.conf
|
||||||
@@ -1,29 +0,0 @@
|
|||||||
- name: Create k3s-folder
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: /etc/rancher/k3s/
|
|
||||||
state: directory
|
|
||||||
mode: '0755'
|
|
||||||
|
|
||||||
- name: Deploy k3s config
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: k3s/{{ type }}/config.yaml.jinja2
|
|
||||||
dest: /etc/rancher/k3s/config.yaml
|
|
||||||
|
|
||||||
- name: Download install-script
|
|
||||||
get_url:
|
|
||||||
url: https://get.k3s.io
|
|
||||||
dest: /root/k3s_install.sh
|
|
||||||
mode: '744'
|
|
||||||
# todo: update when file changed?
|
|
||||||
|
|
||||||
- import_tasks: ./install/server/setup_network.yml
|
|
||||||
when: "type == 'server'"
|
|
||||||
|
|
||||||
- import_tasks: ./install/server/install_helm.yml
|
|
||||||
when: "type == 'server'"
|
|
||||||
|
|
||||||
- import_tasks: ./install/server/install_k3s.yml
|
|
||||||
when: "type == 'server'"
|
|
||||||
|
|
||||||
- import_tasks: ./install/agent/install_k3s.yml
|
|
||||||
when: "type == 'agent'"
|
|
||||||
@@ -1,7 +0,0 @@
|
|||||||
- name: Install K3s agent
|
|
||||||
command: /root/k3s_install.sh {{ type }}
|
|
||||||
register: command
|
|
||||||
changed_when: "'No change detected' in command.stdout"
|
|
||||||
until: "command is not failed"
|
|
||||||
retries: 2
|
|
||||||
delay: 10
|
|
||||||
@@ -1,17 +0,0 @@
|
|||||||
- name: Add Balto key
|
|
||||||
apt_key:
|
|
||||||
url: https://baltocdn.com/helm/signing.asc
|
|
||||||
state: present
|
|
||||||
|
|
||||||
- name: Add Balto Repository
|
|
||||||
apt_repository:
|
|
||||||
repo: "deb https://baltocdn.com/helm/stable/debian/ all main"
|
|
||||||
state: present
|
|
||||||
filename: kubernetes
|
|
||||||
update_cache: yes
|
|
||||||
|
|
||||||
- name: Install helm
|
|
||||||
package:
|
|
||||||
name:
|
|
||||||
- helm
|
|
||||||
state: latest
|
|
||||||
@@ -1,35 +0,0 @@
|
|||||||
- name: Install K3s-server for 1st-node
|
|
||||||
command: /root/k3s_install.sh {{ type }}
|
|
||||||
when: "inventory_hostname == groups['kubernetes'][0]"
|
|
||||||
register: command
|
|
||||||
changed_when: "'No change detected' in command.stdout"
|
|
||||||
|
|
||||||
- name: Waiting for K3s-server to accept connections
|
|
||||||
ansible.builtin.wait_for:
|
|
||||||
host: "{{ inventory_hostname }}"
|
|
||||||
port: 6443
|
|
||||||
state: started
|
|
||||||
when: "inventory_hostname == groups['kubernetes'][0]"
|
|
||||||
|
|
||||||
- name: Install K3s-server for other nodes
|
|
||||||
command: /root/k3s_install.sh {{ type }}
|
|
||||||
when: "inventory_hostname != groups['kubernetes'][0]"
|
|
||||||
register: command
|
|
||||||
changed_when: "'No change detected' in command.stdout"
|
|
||||||
until: "command is not failed"
|
|
||||||
retries: 2
|
|
||||||
delay: 10
|
|
||||||
|
|
||||||
- name: Waiting for K3s-server to accept connections on other nodes
|
|
||||||
ansible.builtin.wait_for:
|
|
||||||
host: "{{ inventory_hostname }}"
|
|
||||||
port: 6443
|
|
||||||
state: started
|
|
||||||
when: "inventory_hostname != groups['kubernetes'][0]"
|
|
||||||
#- name: Add Kubernetes environment-vars to /etc/profile.d/
|
|
||||||
# blockinfile:
|
|
||||||
# path: /etc/profile.d/k3s-bin.sh
|
|
||||||
# marker: "# {mark} ANSIBLE MANAGED BLOCK | k3s"
|
|
||||||
# block: |
|
|
||||||
# export KUBECONFIG="/etc/rancher/k3s/k3s.yaml"
|
|
||||||
# create: true
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
- name: Set control-plane-dns-endpoint towards local-ip
|
|
||||||
blockinfile:
|
|
||||||
path: /etc/hosts
|
|
||||||
marker: "# {mark} ANSIBLE MANAGED BLOCK | k3s"
|
|
||||||
block: |
|
|
||||||
{{ nodeip_ipv4 }} {{ kubernetes.control_plane.dns_name }}
|
|
||||||
28
kubernetes/tasks/k8s_deploy.yml
Normal file
28
kubernetes/tasks/k8s_deploy.yml
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
- name: Add Google-Cloud key
|
||||||
|
apt_key:
|
||||||
|
url: https://packages.cloud.google.com/apt/doc/apt-key.gpg
|
||||||
|
state: present
|
||||||
|
|
||||||
|
- name: Add Kubernetes Repository
|
||||||
|
apt_repository:
|
||||||
|
repo: "deb https://apt.kubernetes.io/ kubernetes-xenial main"
|
||||||
|
state: present
|
||||||
|
filename: kubernetes
|
||||||
|
update_cache: yes
|
||||||
|
|
||||||
|
- name: Install kubernetes-tools
|
||||||
|
package:
|
||||||
|
name:
|
||||||
|
- kubeadm
|
||||||
|
- kubelet
|
||||||
|
- kubectl
|
||||||
|
state: latest
|
||||||
|
|
||||||
|
- name: Hold upgrades for kubernetes-tools
|
||||||
|
dpkg_selections:
|
||||||
|
name: "{{ item }}"
|
||||||
|
selection: hold
|
||||||
|
loop:
|
||||||
|
- kubeadm
|
||||||
|
- kubelet
|
||||||
|
- kubectl
|
||||||
4
kubernetes/tasks/k8s_setup-cluster.yml
Normal file
4
kubernetes/tasks/k8s_setup-cluster.yml
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
- name: Join other nodes to cluster
|
||||||
|
when: "inventory_hostname != groups['kubernetes'][0]"
|
||||||
|
command:
|
||||||
|
|
||||||
9
kubernetes/tasks/k8s_setup.yml
Normal file
9
kubernetes/tasks/k8s_setup.yml
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
- name: Initialize Kubernetes Cluster
|
||||||
|
when: "inventory_hostname == groups['kubernetes'][0]"
|
||||||
|
command: kubeadm init --control-plane-endpoint={{ control_plane.dns_name }}
|
||||||
|
#--upload-certs
|
||||||
|
|
||||||
|
- name: Set environment-var for config
|
||||||
|
lineinfile:
|
||||||
|
dest: ~/.bashrc
|
||||||
|
line: "export KUBECONFIG=/etc/kubernetes/admin.conf"
|
||||||
@@ -1,4 +1,9 @@
|
|||||||
- import_tasks: ./prerequisites.yml
|
- import_tasks: ./prerequisites.yml
|
||||||
|
|
||||||
- import_tasks: ./install.yml
|
- import_tasks: ./k8s_deploy.yml
|
||||||
|
|
||||||
|
- import_tasks: ./k8s_setup.yml
|
||||||
|
|
||||||
|
- import_tasks: ./deploy_cilium.yml
|
||||||
|
|
||||||
|
#- import_tasks: ./k8s_setup-cluster.yml
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
#- name: Load br_netfilter kernel-module
|
- name: Load br_netfilter kernel-module
|
||||||
# modprobe:
|
modprobe:
|
||||||
# name: br_netfilter
|
name: br_netfilter
|
||||||
# state: present
|
state: present
|
||||||
|
|
||||||
- name: Set sysctl settings for iptables bridged traffic
|
- name: Set sysctl settings for iptables bridged traffic
|
||||||
copy:
|
copy:
|
||||||
@@ -9,27 +9,13 @@
|
|||||||
content: |
|
content: |
|
||||||
net.bridge.bridge-nf-call-ip6tables = 1
|
net.bridge.bridge-nf-call-ip6tables = 1
|
||||||
net.bridge.bridge-nf-call-iptables = 1
|
net.bridge.bridge-nf-call-iptables = 1
|
||||||
|
|
||||||
net.ipv4.conf.all.forwarding=1
|
|
||||||
net.ipv6.conf.all.forwarding=1
|
|
||||||
notify: reload_sysctl
|
notify: reload_sysctl
|
||||||
|
|
||||||
#- name: Disable swap
|
- import_tasks: ./prerequisites/swap.yml
|
||||||
# command: swapoff -a
|
|
||||||
|
|
||||||
#- name: Install iptables
|
|
||||||
# package:
|
|
||||||
# name:
|
|
||||||
# #- containerd
|
|
||||||
# - iptables
|
|
||||||
# state: latest
|
|
||||||
|
|
||||||
- import_tasks: ./prerequisites/containerd.yml
|
- import_tasks: ./prerequisites/containerd.yml
|
||||||
|
|
||||||
- name: Getting nodeIp-data from interface
|
- name: Set control-plane-dns-endpoint towards local-ip
|
||||||
set_fact:
|
lineinfile:
|
||||||
nodeip_ipv4: "{{ ansible_facts[ kubernetes.ipPool.nodeIp_interface ].ipv4.address }}"
|
dest: /etc/hosts
|
||||||
nodeip_ipv6: "{{ ansible_facts[ kubernetes.ipPool.nodeIp_interface ].ipv6[0].address }}"
|
line: "{{ ansible_facts.default_ipv6.address }} k8s-control-plane.system.ruekov.eu"
|
||||||
|
|
||||||
- name: Run handlers to reload configurations
|
|
||||||
meta: flush_handlers
|
|
||||||
|
|||||||
10
kubernetes/tasks/prerequisites/swap.yml
Normal file
10
kubernetes/tasks/prerequisites/swap.yml
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
- name: Disable swap-mounts
|
||||||
|
replace:
|
||||||
|
path: /etc/fstab
|
||||||
|
regexp: '^([ \t]*(?!#)\S+[ \t]+swap[ \t]+.*)'
|
||||||
|
replace: '# \1'
|
||||||
|
|
||||||
|
- name: Disable active swap immediately
|
||||||
|
command: swapoff -va
|
||||||
|
changed_when: "command.stdout != ''"
|
||||||
|
register: command
|
||||||
@@ -1,7 +0,0 @@
|
|||||||
server: https://{{ kubernetes.control_plane.dns_name }}:6443
|
|
||||||
token: {{ kubernetes.token }}
|
|
||||||
|
|
||||||
node-ip: {{ nodeip_ipv4 }},{{ nodeip_ipv6 }}
|
|
||||||
|
|
||||||
# FIXME: Workaround for bug in Kubernetes 1.24/1.25 ignoring node IPv6 addresses
|
|
||||||
kubelet-arg: "--node-ip=0.0.0.0"
|
|
||||||
@@ -1,23 +0,0 @@
|
|||||||
## Base ##
|
|
||||||
{% if inventory_hostname == groups['kubernetes'][0] %}
|
|
||||||
cluster-init: true
|
|
||||||
{% else %}
|
|
||||||
server: https://{{ groups['kubernetes'][0] }}:6443
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
token: {{ kubernetes.token }}
|
|
||||||
tls-san:
|
|
||||||
- {{ kubernetes.control_plane.dns_name }}
|
|
||||||
|
|
||||||
# Networking
|
|
||||||
node-ip: {{ nodeip_ipv4 }},{{ nodeip_ipv6 }}
|
|
||||||
cluster-cidr: {{ kubernetes.ipPool.ipv4.cluster_cidr }},{{ kubernetes.ipPool.ipv6.cluster_cidr }}
|
|
||||||
service-cidr: {{ kubernetes.ipPool.ipv4.service_cidr }},{{ kubernetes.ipPool.ipv6.service_cidr }}
|
|
||||||
|
|
||||||
egress-selector-mode: disabled
|
|
||||||
|
|
||||||
# Network-plugin
|
|
||||||
flannel-backend: vxlan
|
|
||||||
|
|
||||||
# FIXME: Workaround for bug in Kubernetes 1.24/1.25 ignoring node IPv6 addresses
|
|
||||||
kubelet-arg: "--node-ip=0.0.0.0"
|
|
||||||
7
kubernetes/templates/k8s-dqlite/init.yaml.jinja2
Normal file
7
kubernetes/templates/k8s-dqlite/init.yaml.jinja2
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
{% if inventory_hostname != groups['kubernetes'][0] %}
|
||||||
|
Cluster:
|
||||||
|
{% for node in groups['kubernetes'] if node != inventory_hostname %}
|
||||||
|
- {{ node }}:29001
|
||||||
|
{% endfor %}
|
||||||
|
{% endif %}
|
||||||
|
Address: 0.0.0.0:29001
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
---
|
|
||||||
nomad:
|
|
||||||
version: 1.4.4
|
|
||||||
@@ -1,48 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=Nomad
|
|
||||||
Documentation=https://www.nomadproject.io/docs/
|
|
||||||
Wants=network-online.target
|
|
||||||
After=network-online.target
|
|
||||||
|
|
||||||
# When using Nomad with Consul it is not necessary to start Consul first. These
|
|
||||||
# lines start Consul before Nomad as an optimization to avoid Nomad logging
|
|
||||||
# that Consul is unavailable at startup.
|
|
||||||
#Wants=consul.service
|
|
||||||
#After=consul.service
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
|
|
||||||
# Nomad server should be run as the nomad user. Nomad clients
|
|
||||||
# should be run as root
|
|
||||||
User=root
|
|
||||||
Group=root
|
|
||||||
|
|
||||||
ExecReload=/bin/kill -HUP $MAINPID
|
|
||||||
ExecStart=/usr/local/bin/nomad agent -config /etc/nomad.d
|
|
||||||
KillMode=process
|
|
||||||
KillSignal=SIGINT
|
|
||||||
LimitNOFILE=65536
|
|
||||||
LimitNPROC=infinity
|
|
||||||
Restart=on-failure
|
|
||||||
RestartSec=2
|
|
||||||
|
|
||||||
## Configure unit start rate limiting. Units which are started more than
|
|
||||||
## *burst* times within an *interval* time span are not permitted to start any
|
|
||||||
## more. Use `StartLimitIntervalSec` or `StartLimitInterval` (depending on
|
|
||||||
## systemd version) to configure the checking interval and `StartLimitBurst`
|
|
||||||
## to configure how many starts per interval are allowed. The values in the
|
|
||||||
## commented lines are defaults.
|
|
||||||
|
|
||||||
# StartLimitBurst = 5
|
|
||||||
|
|
||||||
## StartLimitIntervalSec is used for systemd versions >= 230
|
|
||||||
# StartLimitIntervalSec = 10s
|
|
||||||
|
|
||||||
## StartLimitInterval is used for systemd versions < 230
|
|
||||||
# StartLimitInterval = 10s
|
|
||||||
|
|
||||||
TasksMax=infinity
|
|
||||||
OOMScoreAdjust=-1000
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
---
|
|
||||||
dependencies:
|
|
||||||
#- role: docker
|
|
||||||
@@ -1,43 +0,0 @@
|
|||||||
- name: Download binary
|
|
||||||
ansible.builtin.unarchive:
|
|
||||||
remote_src: true
|
|
||||||
src: https://releases.hashicorp.com/nomad/{{ nomad.version }}/nomad_{{ nomad.version }}_{{ ansible_system | lower }}_{{ 'amd64' if ansible_architecture == 'x86_64' else ansible_architecture }}.zip
|
|
||||||
dest: /usr/local/bin/
|
|
||||||
mode: "755"
|
|
||||||
|
|
||||||
- name: Deploy systemd-service file
|
|
||||||
ansible.builtin.copy:
|
|
||||||
src: systemd-service
|
|
||||||
dest: /etc/systemd/system/nomad.service
|
|
||||||
mode: u=rw,g=r,o=r
|
|
||||||
|
|
||||||
- name: Create nomad user
|
|
||||||
ansible.builtin.user:
|
|
||||||
name: nomad
|
|
||||||
groups:
|
|
||||||
- docker
|
|
||||||
append: true
|
|
||||||
|
|
||||||
- name: Create directory for configs
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: /etc/nomad.d
|
|
||||||
state: directory
|
|
||||||
mode: "0755"
|
|
||||||
owner: "nomad"
|
|
||||||
group: "nomad"
|
|
||||||
|
|
||||||
- name: Create nomad.hcl configuration file
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: nomad.hcl.j2
|
|
||||||
dest: /etc/nomad.d/nomad.hcl
|
|
||||||
mode: "0644"
|
|
||||||
owner: "nomad"
|
|
||||||
group: "nomad"
|
|
||||||
|
|
||||||
- name: Create directory for data
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: /opt/nomad
|
|
||||||
state: directory
|
|
||||||
mode: "0755"
|
|
||||||
owner: "nomad"
|
|
||||||
group: "nomad"
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
- name: Start service
|
|
||||||
ansible.builtin.service:
|
|
||||||
name: nomad
|
|
||||||
state: restarted
|
|
||||||
|
|
||||||
- name: Waiting for service to accept connections
|
|
||||||
ansible.builtin.wait_for:
|
|
||||||
port: 4646
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
- import_tasks: ./install.yml
|
|
||||||
|
|
||||||
- import_tasks: ./launch.yml
|
|
||||||
@@ -1,71 +0,0 @@
|
|||||||
data_dir = "/opt/nomad"
|
|
||||||
datacenter = "{{ datacenter }}"
|
|
||||||
|
|
||||||
|
|
||||||
bind_addr = "0.0.0.0"
|
|
||||||
|
|
||||||
advertise {
|
|
||||||
# Defaults to the first private IP address.
|
|
||||||
#http = "1.2.3.4"
|
|
||||||
#rpc = "1.2.3.4"
|
|
||||||
#serf = "1.2.3.4:5648" # non-default ports may be specified
|
|
||||||
}
|
|
||||||
|
|
||||||
{# TODO: Get interface-ip from hosts marked with type=server #}
|
|
||||||
{% set server_hosts = ansible_play_batch | difference([inventory_hostname]) %}
|
|
||||||
{% if type is defined and type == "server" %}
|
|
||||||
server {
|
|
||||||
enabled = true
|
|
||||||
bootstrap_expect = {{ server_hosts | length }}
|
|
||||||
|
|
||||||
server_join {
|
|
||||||
retry_join = [ "{{ server_hosts | join('", "') }}" ]
|
|
||||||
retry_max = 6
|
|
||||||
retry_interval = "15s"
|
|
||||||
}
|
|
||||||
|
|
||||||
default_scheduler_config {
|
|
||||||
scheduler_algorithm = "binpack"
|
|
||||||
memory_oversubscription_enabled = true
|
|
||||||
reject_job_registration = false
|
|
||||||
pause_eval_broker = false # New in Nomad 1.3.2
|
|
||||||
|
|
||||||
preemption_config {
|
|
||||||
batch_scheduler_enabled = true
|
|
||||||
system_scheduler_enabled = true
|
|
||||||
service_scheduler_enabled = true
|
|
||||||
sysbatch_scheduler_enabled = true # New in Nomad 1.2
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
client {
|
|
||||||
enabled = true
|
|
||||||
|
|
||||||
{% if type != "server" %}
|
|
||||||
servers = [ "{{ server_hosts | join('", "') }}" ]
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
meta {
|
|
||||||
node_type = "{{ type }}"
|
|
||||||
{% if storage is defined and storage %}
|
|
||||||
seaweedfs_volume = "true"
|
|
||||||
{% endif %}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
plugin "raw_exec" {
|
|
||||||
config {
|
|
||||||
enabled = true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
plugin "docker" {
|
|
||||||
config {
|
|
||||||
{% if type is defined and type == "server" %}
|
|
||||||
allow_privileged = true
|
|
||||||
{% endif %}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Reference in New Issue
Block a user