From b2d250cf85469d7e1b0a8ad42634c04399b98439 Mon Sep 17 00:00:00 2001 From: Thorsten Bruhns Date: Fri, 16 Sep 2022 08:29:19 +0000 Subject: [PATCH 1/5] inventory: Example for Oracle Restart (has) --- README.adoc | 38 +++++++- .../dbfs/group_vars/all/software_src.yml | 2 +- inventory/has/group_vars/all/asm.yml | 25 +++++ inventory/has/group_vars/all/database.yml | 96 +++++++++++++++++++ inventory/has/group_vars/all/dev-sec.yml | 18 ++++ inventory/has/group_vars/all/host.yml | 28 ++++++ inventory/has/group_vars/all/software_src.yml | 13 +++ inventory/has/hosts.yml | 8 ++ 8 files changed, 226 insertions(+), 2 deletions(-) create mode 100644 inventory/has/group_vars/all/asm.yml create mode 100644 inventory/has/group_vars/all/database.yml create mode 100644 inventory/has/group_vars/all/dev-sec.yml create mode 100644 inventory/has/group_vars/all/host.yml create mode 100755 inventory/has/group_vars/all/software_src.yml create mode 100755 inventory/has/hosts.yml diff --git a/README.adoc b/README.adoc index 9f21b442c..ca64dabbf 100644 --- a/README.adoc +++ b/README.adoc @@ -64,7 +64,7 @@ cd vagrant/dbfs VAGRANT_EXPERIMENTAL=disks vagrant up ---- -IMPORTANT: Copy `LINUX.X64_190000_db_home.zip` into `/tmp` on Vagrantbox +IMPORTANT: Copy `LINUX.X64_193000_grid_home.zip` into `/vagrant` in same directory as the Vagrantfile. #### Start Playbook @@ -77,6 +77,42 @@ The `playbook/os_vagrant.yml` fixes that. ansible-playbook -e hostgroup=all -i inventory/dbfs/hosts.yml playbooks/os_vagrant.yml playbooks/single-instance-fs.yml ---- +### Single Instance with Oracle Restart (HAS) in Vagrantbox + + +#### Vagrantbox for Database Server + +Requirements:: +8500 MB RAM: + +20 GB Diskspace: + +The Inventory from has expects a 2nd disk at `/dev/sdb`. +See `host_fs_layout` as example. + +The Vagrantfile is in `vagrant/has` of `ansible-oracle`. + +IMPORTANT: The `VAGRANT_EXPERIMENTAL=disks` is very important. +It adds the 2nd disk for LVM and 2 ASM-Disks to the VM during the startup. + +.Start Vagrantbox +---- +cd vagrant/has +VAGRANT_EXPERIMENTAL=disks vagrant up +---- + +IMPORTANT: Copy `LINUX.X64_193000_grid_home.zip` and `LINUX.X64_190000_db_home.zip` into `/vagrant` in same directory as the Vagrantfile. + + +#### Start Playbook + +IMPORTANT: The used Vagrantbox does not configure the `/etc/hosts` for Oracle correctly. +The `playbook/os_vagrant.yml` fixes that. + + +.execute playbook +---- +ansible-playbook -e hostgroup=all -i inventory/has/hosts.yml playbooks/os_vagrant.yml playbooks/single-instance-asm.yml +---- + = Roles == common diff --git a/inventory/dbfs/group_vars/all/software_src.yml b/inventory/dbfs/group_vars/all/software_src.yml index 8798876b5..d55b2e115 100755 --- a/inventory/dbfs/group_vars/all/software_src.yml +++ b/inventory/dbfs/group_vars/all/software_src.yml @@ -4,7 +4,7 @@ oracle_sw_copy: false oracle_sw_unpack: false # Directory for Installation-Media -oracle_stage_remote: /tmp +oracle_stage_remote: /vagrant # Example for Remote NFS # install_from_nfs: true # Mount NFS-Share? diff --git a/inventory/has/group_vars/all/asm.yml b/inventory/has/group_vars/all/asm.yml new file mode 100644 index 000000000..ad5e33a45 --- /dev/null +++ b/inventory/has/group_vars/all/asm.yml @@ -0,0 +1,25 @@ +--- +oracle_install_version_gi: 19.3.0.0 +apply_patches_gi: false + +oracle_asm_init_dg: data # 1st diskgroup + +asm_diskgroups: # ASM Diskgroups used for DB-storage. Should map to dict asm_storage_layout. + - diskgroup: data + state: present + properties: + - {redundancy: external, ausize: 4} + attributes: + - {name: compatible.rdbms, value: "19.0.0.0.0"} + - {name: compatible.asm, value: "19.0.0.0.0"} + disk: + - {device: /dev/sdc, asmlabel: data01} + - diskgroup: fra + state: present + properties: + - {redundancy: external, ausize: 4} + attributes: + - {name: compatible.rdbms, value: "19.0.0.0.0"} + - {name: compatible.asm, value: "19.0.0.0.0"} + disk: + - {device: /dev/sdd, asmlabel: fra01} diff --git a/inventory/has/group_vars/all/database.yml b/inventory/has/group_vars/all/database.yml new file mode 100644 index 000000000..059e1dc48 --- /dev/null +++ b/inventory/has/group_vars/all/database.yml @@ -0,0 +1,96 @@ +--- +# This is an example for 1 Instance on 1 Host. +# Please look at ansible-oracle-inventory for more complex configurations: +# https://github.com/opitzconsulting/ansible-oracle-inventory +# +oracle_databases: + - home: &db_config_home 19300-base + oracle_db_name: &oracle_db_name ORCL + oracle_db_type: SI # Type of database (RAC,RACONENODE,SI) + is_container: true + storage_type: ASM # Database storage to be used. ASM or FS. + oracle_db_mem_totalmb: 1536 # Amount of RAM to be used for SGA + oracle_database_type: MULTIPURPOSE # MULTIPURPOSE|DATA_WAREHOUSING|OLTP + redolog_size: 75M + redolog_groups: 3 + datafile_dest: +DATA + recoveryfile_dest: +FRA + # listener_name: LISTENER # This home will have a listener configured + listener_port: &cdb_listener_port 1521 + # *local_listener is used in initparam as an achor + local_listener: &local_listener "'{{ ansible_hostname }}:1521'" + archivelog: false + flashback: false + force_logging: true + state: present + statspack: + purgedays: 14 + snaplevel: 5 + state: present + tablespaces: + - {name: system, size: 10M, autoextend: true, next: 50M, maxsize: 4G, content: permanent, state: present, bigfile: false} + - {name: sysaux, size: 10M, autoextend: true, next: 50M, maxsize: 4G, content: permanent, state: present, bigfile: false} + - {name: undotbs1, size: 10M, autoextend: true, next: 50M, maxsize: 8G, content: permanent, state: present, bigfile: false} + - {name: users, size: 10M, autoextend: true, next: 50M, maxsize: 2G, content: permanent, state: present, bigfile: false} + - {name: temp, size: 10M, autoextend: true, next: 50M, maxsize: 4G, content: permanent, state: present, bigfile: false} + init_parameters: + - {name: audit_trail, value: 'NONE', scope: spfile, state: present} + - {name: processes, value: '400', scope: spfile, state: present, dbca: false} + # - {name: local_listener, value: *local_listener, scope: both, state: present} + - {name: archive_lag_target, value: '900', scope: both, state: present} + - {name: control_management_pack_access, value: 'NONE', scope: both, state: present} + - {name: control_file_record_keep_time, value: '30', scope: both, state: present} + - {name: db_files, value: '200', scope: spfile, state: present} + - {name: deferred_segment_creation, value: 'false', scope: both, state: present} + - {name: filesystemio_options, value: 'setall', scope: spfile, state: present} + - {name: job_queue_processes, value: '10', scope: both, state: present} + # Disable forcing hugepages on really small systems + # - {name: use_large_pages ,value: 'ONLY', scope: spfile, state: present} + - {name: log_archive_dest_1, value: 'location=USE_DB_RECOVERY_FILE_DEST', scope: both, state: present} + - {name: log_buffer, value: '64M', scope: spfile, state: present} + - {name: pga_aggregate_target, value: '200M', scope: both, state: present, dbca: false} + - {name: sga_target, value: '1800M', scope: spfile, state: present, dbca: false} + - {name: shared_pool_size, value: '768M', scope: both, state: present, dbca: true} + - {name: recyclebin, value: 'off', scope: spfile, state: present} + - {name: standby_file_management, value: 'AUTO', scope: both, state: present} + - {name: streams_pool_size, value: '152M', scope: spfile, state: present} + - {name: "_cursor_obsolete_threshold", value: '1024', scope: spfile, state: present} + - {name: max_pdbs, value: '3', scope: both, state: present} + - {name: clonedb, value: 'true', scope: spfile, state: present, dbca: false} + - {name: db_create_file_dest, value: '+DATA', scope: both, state: present} + - {name: db_create_online_log_dest_1, value: '+DATA', scope: both, state: present} + - {name: db_recovery_file_dest_size, value: '10G', scope: both, state: present, dbca: false} + + profiles: + - name: DEFAULT + state: present + attributes: + - {name: password_life_time, value: unlimited} + + users: + - schema: dbsnmp + state: unlocked + update_password: always + + rman_jobs: + - {name: parameter} + - {name: offline_level0, disabled: false, weekday: "0", hour: "01", minute: "10", day: "*"} + +oracle_pdbs: + - home: *db_config_home + listener_port: *cdb_listener_port + cdb: *oracle_db_name + pdb_name: ORCLPDB + state: present + profiles: "{{ oracle_default_profiles }}" + statspack: + purgedays: 14 + snaplevel: 7 + state: present + + tablespaces: + - {name: system, size: 10M, autoextend: true, next: 50M, maxsize: 4G, content: permanent, state: present, bigfile: false} + - {name: sysaux, size: 10M, autoextend: true, next: 50M, maxsize: 4G, content: permanent, state: present, bigfile: false} + - {name: undotbs1, size: 10M, autoextend: true, next: 50M, maxsize: 8G, content: permanent, state: present, bigfile: false} + - {name: users, size: 10M, autoextend: true, next: 50M, maxsize: 2G, content: permanent, state: present, bigfile: false} + - {name: temp, size: 10M, autoextend: true, next: 50M, maxsize: 4G, content: permanent, state: present, bigfile: false} diff --git a/inventory/has/group_vars/all/dev-sec.yml b/inventory/has/group_vars/all/dev-sec.yml new file mode 100644 index 000000000..3f98f8b80 --- /dev/null +++ b/inventory/has/group_vars/all/dev-sec.yml @@ -0,0 +1,18 @@ +--- +# Oracle gets problems, when root processes are not visible +hidepid_option: 0 + +os_security_kernel_enable_module_loading: true + +sysctl_overwrite: + network_ipv6_enable: false + fs.protected_regular: 0 # needed for opatchauto ... + +# ssh settings +ssh_print_last_log: true +ssh_allow_agent_forwarding: false +ssh_permit_tunnel: false +ssh_allow_tcp_forwarding: 'no' +ssh_max_auth_retries: 3 + +ssh_allow_users: vagrant ansible diff --git a/inventory/has/group_vars/all/host.yml b/inventory/has/group_vars/all/host.yml new file mode 100644 index 000000000..ddd6ab1c0 --- /dev/null +++ b/inventory/has/group_vars/all/host.yml @@ -0,0 +1,28 @@ +--- +configure_public_yum_repo: false +configure_motd: false + +configure_hugepages_by: memory + +# disable hugepages on small systems +# don't forget to enable use_large_pages in oracle parameter +# size_in_gb_hugepages: 2 +size_in_gb_hugepages: 0 + + +configure_host_disks: true + +host_fs_layout: + - vgname: vgora + state: present + filesystem: + - {mntp: /u01, lvname: orabaselv, lvsize: 40G, fstype: xfs} + - {mntp: swap, lvname: swaplv, lvsize: 16g, fstype: swap} + disk: + - {device: /dev/sdb, pvname: /dev/sdb1} + - vgname: rootvg + state: present + filesystem: + - {mntp: /tmp, lvname: tmplv, lvsize: 1200m, fstype: ext4} + disk: + - {device: /dev/sda, pvname: /dev/sda2} diff --git a/inventory/has/group_vars/all/software_src.yml b/inventory/has/group_vars/all/software_src.yml new file mode 100755 index 000000000..d55b2e115 --- /dev/null +++ b/inventory/has/group_vars/all/software_src.yml @@ -0,0 +1,13 @@ +--- +is_sw_source_local: true +oracle_sw_copy: false +oracle_sw_unpack: false + +# Directory for Installation-Media +oracle_stage_remote: /vagrant + +# Example for Remote NFS +# install_from_nfs: true # Mount NFS-Share? +# nfs_server_sw: 192.168.56.99 # NFS-Server +# nfs_server_sw_path: /sw # NFS-Share +# oracle_stage_remote: /u01/se # local mount point for NFS share diff --git a/inventory/has/hosts.yml b/inventory/has/hosts.yml new file mode 100755 index 000000000..cfa056f83 --- /dev/null +++ b/inventory/has/hosts.yml @@ -0,0 +1,8 @@ +--- +all: + children: + has: + hosts: + has-192-168-56-162.nip.io: + ansible_connection: ssh + ansible_ssh_user: vagrant From 678fbcfe78cf4466e549edb706187948d3d5ed91 Mon Sep 17 00:00:00 2001 From: Thorsten Bruhns Date: Fri, 16 Sep 2022 10:34:55 +0200 Subject: [PATCH 2/5] vagrant: Vagrantfile for has --- .gitignore | 5 ++ vagrant/dbfs/Vagrantfile | 95 ++++++++++++++++++++++++++++++++++++ vagrant/has/Vagrantfile | 101 +++++++++++++++++++++++++++++++++++++++ 3 files changed, 201 insertions(+) create mode 100644 .gitignore create mode 100644 vagrant/dbfs/Vagrantfile create mode 100644 vagrant/has/Vagrantfile diff --git a/.gitignore b/.gitignore new file mode 100644 index 000000000..6b502878d --- /dev/null +++ b/.gitignore @@ -0,0 +1,5 @@ + +vagrant/*/.vagrant +vagrant/*/*pub +vagrant/*/*zip +vagrant/*/*ZIP diff --git a/vagrant/dbfs/Vagrantfile b/vagrant/dbfs/Vagrantfile new file mode 100644 index 000000000..075155c71 --- /dev/null +++ b/vagrant/dbfs/Vagrantfile @@ -0,0 +1,95 @@ +# -*- mode: ruby -*- +# vi: set ft=ruby : +# +# PLEASE START 1st with following command to create additional disks! +# +# VAGRANT_EXPERIMENTAL=disks vagrant up + +Vagrant.configure("2") do |config| + + lanip = "192.168.56.161" + vm_name = "dbfs161-192-168-56-161" + + puts " " + puts " Single Instance Database in Filesystem" + puts " " + puts " Hostname: #{vm_name}" + puts " IP: #{lanip}" + puts " Logins:" + puts " vagrant / vagrant" + puts " root / vagrant" + puts " " + puts " Please do the 1st start with following command to create the additional disks!" + puts " VAGRANT_EXPERIMENTAL=disks vagrant up" + puts " " + + # Please define the folder to the extracted archives from Oracle + # + config.vm.synced_folder ENV['VAGRANT_ANSIBLE_ORACLE_SW'], "/sw/oracle", disabled: false, mount_options: ["ro"] + + config.vm.disk :disk, size: "150GB", name: "#{vm_name}_oracle" + + # config.disksize.size = '50GB' + config.vm.box = "Rendanic/oraclelinux-7.x" + + config.vm.box_check_update = false + + config.vm.network :private_network, ip: lanip + + # config.vm.network "public_network" + config.ssh.insert_key = true + config.ssh.username = "vagrant" + config.ssh.password = "vagrant" + + config.vm.hostname = "#{vm_name}.nio.io" + + config.vm.provider "virtualbox" do |vb| + vb.name = vm_name + vb.gui = false + vb.memory = "4096" + vb.cpus = 2 + # vb.customize ["modifyvm", :id, "--groups", "/ansible-oracle"] + + end + + config.vm.provision "shell", inline: <<-SHELL + + if [ ! -b /dev/sdb ]; then + echo "Please start the Box with following command once to create the addional disks!" + echo "VAGRANT_EXPERIMENTAL=disks vagrant up" + exit 1 + fi + + # copy public key to vagrant user + ssh_public_key=/vagrant/id_rsa.pub + if test -f "$ssh_public_key" ; then + echo "Check if public key $ssh_public_key is valid." + if ssh-keygen -l -f "$ssh_public_key" ; then + echo "Add public key to /home/vagrant/.ssh/authorized_keys" + cat "$ssh_public_key" >> /home/vagrant/.ssh/authorized_keys + fi + fi + + sed -i 's/^PasswordAuthentication no/PasswordAuthentication yes/g' /etc/ssh/sshd_config + systemctl restart sshd.service + + set -e + # configure local Nexus-Server when running + # + os_release=$(grep '^VERSION_ID' /etc/os-release | awk -F'=' ' gsub(/"/,"") { print $2}' | cut -b1) + if curl -f -s -q -m 3 http://192.168.56.12:8081/repository/ol${os_release}os-yum-group/ > /dev/null ; then + echo "Local Nexus3 Server for yum proxy found!" + + yum-config-manager --disable '*' > /dev/null + yum-config-manager -q --nogpgcheck --add-repo http://192.168.56.12:8081/repository/ol${os_release}os-yum-group + yum-config-manager -q --add-repo http://192.168.56.12:8081/repository/rhel${os_release}_epel-yum-proxy + sed -i 's/^gpgcheck=1/gpgcheck=0/g' /etc/yum.conf + echo "timeout=900" >> /etc/yum.conf + fi + + yum install -y gcc make perl kernel-uek-devel + yum update -y + /sbin/rcvboxadd quicksetup all + + SHELL +end diff --git a/vagrant/has/Vagrantfile b/vagrant/has/Vagrantfile new file mode 100644 index 000000000..05a121038 --- /dev/null +++ b/vagrant/has/Vagrantfile @@ -0,0 +1,101 @@ +# -*- mode: ruby -*- +# vi: set ft=ruby : +# +# PLEASE START 1st with following command to create additional disks! +# +# VAGRANT_EXPERIMENTAL=disks vagrant up + +Vagrant.configure("2") do |config| + + lanip = "192.168.56.162" + vm_name = "has-162" + vm_hostname = "has-192-168-56-162.nip.io" + + # Please define the folder to the extracted archives from Oracle + # + config.vm.synced_folder ENV['VAGRANT_ANSIBLE_ORACLE_SW'], "/sw/oracle", disabled: false, mount_options: ["ro"] + + config.vm.disk :disk, size: "150GB", name: "#{vm_name}_oracle" + config.vm.disk :disk, size: "30GB", name: "#{vm_name}_asmdata1" + config.vm.disk :disk, size: "30GB", name: "#{vm_name}_asmfra1" + + # config.disksize.size = '50GB' + config.vm.box = "Rendanic/oraclelinux-7.x" + + puts " " + puts " Oracle Restart" + puts " " + puts " Hostname: #{vm_hostname}" + puts " IP: #{lanip}" + puts " Logins:" + puts " vagrant / vagrant" + puts " root / vagrant" + puts " " + puts " Please do the 1st start with following command to create the additional disks!" + puts " VAGRANT_EXPERIMENTAL=disks vagrant up" + puts " " + + config.vm.box_check_update = true + + config.vm.network :private_network, ip: lanip + + # config.vm.network "public_network" + config.ssh.insert_key = true + config.ssh.username = "vagrant" + config.ssh.password = "vagrant" + + config.vm.hostname = vm_hostname + + config.vm.provider "virtualbox" do |vb| + vb.gui = false + vb.memory = "8500" + vb.cpus = 2 + vb.name = vm_name + # grouping in VirtualBox is sometimes not working... + # => disable it by default + # vb.customize ["modifyvm", :id, "--groups", "/ansible-oracle"] + + end + + + config.vm.provision "shell", inline: <<-SHELL + + if [ ! -b /dev/sdb ]; then + echo "Please start the Box with following command once to create the addional disks!" + echo "VAGRANT_EXPERIMENTAL=disks vagrant up" + exit 1 + fi + + sed -i 's/^PasswordAuthentication no/PasswordAuthentication yes/g' /etc/ssh/sshd_config + systemctl restart sshd.service + + set -e + # configure local Nexus-Server when running + # + os_release=$(grep '^VERSION_ID' /etc/os-release | awk -F'=' ' gsub(/"/,"") { print $2}' | cut -b1) + if curl -f -s -q -m 3 http://192.168.56.12:8081/repository/ol${os_release}os-yum-group/ > /dev/null ; then + echo "Local Nexus3 Server for yum proxy found!" + + yum-config-manager --disable '*' > /dev/null + yum-config-manager -q --nogpgcheck --add-repo http://192.168.56.12:8081/repository/ol${os_release}os-yum-group + yum-config-manager -q --add-repo http://192.168.56.12:8081/repository/rhel${os_release}_epel-yum-proxy + sed -i 's/^gpgcheck=1/gpgcheck=0/g' /etc/yum.conf + echo "timeout=900" >> /etc/yum.conf + fi + + # copy public key to vagrant user + ssh_public_key=/vagrant/id_rsa.pub + if test -f "$ssh_public_key" ; then + echo "Check if public key $ssh_public_key is valid." + if ssh-keygen -l -f "$ssh_public_key" ; then + echo "Add public key to /home/vagrant/.ssh/authorized_keys" + cat "$ssh_public_key" >> /home/vagrant/.ssh/authorized_keys + fi + fi + + yum install -y gcc make perl kernel-uek-devel + yum update -y + /sbin/rcvboxadd quicksetup all + + SHELL +end From 94c716d312b7d2cdc163fc56d9ba3faa28b06e2a Mon Sep 17 00:00:00 2001 From: Thorsten Bruhns Date: Fri, 16 Sep 2022 11:39:21 +0200 Subject: [PATCH 3/5] changelog --- changelogs/fragments/272-vagrant.yml | 4 ++++ 1 file changed, 4 insertions(+) create mode 100644 changelogs/fragments/272-vagrant.yml diff --git a/changelogs/fragments/272-vagrant.yml b/changelogs/fragments/272-vagrant.yml new file mode 100644 index 000000000..a8ff06ae9 --- /dev/null +++ b/changelogs/fragments/272-vagrant.yml @@ -0,0 +1,4 @@ +--- +minor_changes: + - "vagrant: Vagrantfile for dbfs & has (#272)" + - "inventory: New Inventory for has (#272)" From 7c5e52c6bf048eb7666d218ae626bcb2bc6a40f1 Mon Sep 17 00:00:00 2001 From: Thorsten Bruhns Date: Sun, 2 Oct 2022 17:16:55 +0000 Subject: [PATCH 4/5] desupported leftover racattackl-install.yml --- changelogs/fragments/272-vagrant.yml | 2 ++ playbooks/{ => desupported}/racattackl-install.yml | 0 2 files changed, 2 insertions(+) rename playbooks/{ => desupported}/racattackl-install.yml (100%) diff --git a/changelogs/fragments/272-vagrant.yml b/changelogs/fragments/272-vagrant.yml index a8ff06ae9..cb4b34934 100644 --- a/changelogs/fragments/272-vagrant.yml +++ b/changelogs/fragments/272-vagrant.yml @@ -2,3 +2,5 @@ minor_changes: - "vagrant: Vagrantfile for dbfs & has (#272)" - "inventory: New Inventory for has (#272)" +removed_features: + - "desupported leftover racattackl-install.yml (#272)" diff --git a/playbooks/racattackl-install.yml b/playbooks/desupported/racattackl-install.yml similarity index 100% rename from playbooks/racattackl-install.yml rename to playbooks/desupported/racattackl-install.yml From 2abb0a9e40e211724792c812a91cd0ed0396f096 Mon Sep 17 00:00:00 2001 From: Thorsten Bruhns Date: Sun, 2 Oct 2022 17:24:46 +0000 Subject: [PATCH 5/5] ansible-lint: removed name[play] from execptions --- .ansible-lint | 1 - changelogs/fragments/272-vagrant.yml | 2 ++ playbooks/single-instance-asm.yml | 18 ++++++++++++------ playbooks/single-instance-fs.yml | 15 ++++++++++----- 4 files changed, 24 insertions(+), 12 deletions(-) diff --git a/.ansible-lint b/.ansible-lint index 03b3e4330..e2ad2019e 100644 --- a/.ansible-lint +++ b/.ansible-lint @@ -3,7 +3,6 @@ use_default_rules: true skip_list: - name[casing] - - name[play] exclude_paths: - .github diff --git a/changelogs/fragments/272-vagrant.yml b/changelogs/fragments/272-vagrant.yml index cb4b34934..32c040f36 100644 --- a/changelogs/fragments/272-vagrant.yml +++ b/changelogs/fragments/272-vagrant.yml @@ -4,3 +4,5 @@ minor_changes: - "inventory: New Inventory for has (#272)" removed_features: - "desupported leftover racattackl-install.yml (#272)" +bugfixes: + - "ansible-lint: removed name[play] from execptions (#272)" diff --git a/playbooks/single-instance-asm.yml b/playbooks/single-instance-asm.yml index 1984d5dbc..d6ff3b45e 100644 --- a/playbooks/single-instance-asm.yml +++ b/playbooks/single-instance-asm.yml @@ -1,7 +1,13 @@ --- -- import_playbook: dev-sec.yml -- import_playbook: os.yml -- import_playbook: sql-zauberkastern.yml -- import_playbook: swgi.yml -- import_playbook: swdb.yml -- import_playbook: manage-db.yml +- name: Import_playbook dev-sec + import_playbook: dev-sec.yml +- name: Import_playbook os + import_playbook: os.yml +- name: Import_playbook sql-zauberkastern + import_playbook: sql-zauberkastern.yml +- name: Import_playbook swgi + import_playbook: swgi.yml +- name: Import_playbook swdb + import_playbook: swdb.yml +- name: Import_playbook manage-db + import_playbook: manage-db.yml diff --git a/playbooks/single-instance-fs.yml b/playbooks/single-instance-fs.yml index e608dfee7..e4a50828c 100644 --- a/playbooks/single-instance-fs.yml +++ b/playbooks/single-instance-fs.yml @@ -1,6 +1,11 @@ --- -- import_playbook: dev-sec.yml -- import_playbook: os.yml -- import_playbook: sql-zauberkastern.yml -- import_playbook: swdb.yml -- import_playbook: manage-db.yml +- name: Import_playbook dev-sec + import_playbook: dev-sec.yml +- name: Import_playbook os + import_playbook: os.yml +- name: Import_playbook sql-zauberkastern + import_playbook: sql-zauberkastern.yml +- name: Import_playbook swdb + import_playbook: swdb.yml +- name: Import_playbook manage-db + import_playbook: manage-db.yml